var/home/core/zuul-output/0000755000175000017500000000000015072203307014524 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015072213740015472 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004660371615072213732017714 0ustar rootrootOct 10 13:18:04 crc systemd[1]: Starting Kubernetes Kubelet... Oct 10 13:18:05 crc restorecon[4671]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 13:18:05 crc restorecon[4671]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 13:18:05 crc restorecon[4671]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 10 13:18:06 crc kubenswrapper[4745]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 10 13:18:06 crc kubenswrapper[4745]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 10 13:18:06 crc kubenswrapper[4745]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 10 13:18:06 crc kubenswrapper[4745]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 10 13:18:06 crc kubenswrapper[4745]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 10 13:18:06 crc kubenswrapper[4745]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.513832 4745 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525845 4745 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525884 4745 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525892 4745 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525899 4745 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525906 4745 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525912 4745 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525918 4745 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525924 4745 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525930 4745 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525936 4745 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525942 4745 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525948 4745 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525953 4745 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525958 4745 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525964 4745 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525969 4745 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525974 4745 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525980 4745 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525985 4745 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525990 4745 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.525996 4745 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526001 4745 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526006 4745 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526014 4745 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526022 4745 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526029 4745 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526035 4745 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526040 4745 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526045 4745 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526051 4745 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526064 4745 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526071 4745 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526076 4745 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526082 4745 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526087 4745 feature_gate.go:330] unrecognized feature gate: Example Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526093 4745 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526098 4745 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526103 4745 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526108 4745 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526113 4745 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526119 4745 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526124 4745 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526129 4745 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526135 4745 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526141 4745 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526146 4745 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526152 4745 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526157 4745 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526162 4745 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526168 4745 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526173 4745 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526178 4745 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526183 4745 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526190 4745 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526195 4745 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526200 4745 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526206 4745 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526211 4745 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526217 4745 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526223 4745 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526230 4745 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526237 4745 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526244 4745 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526251 4745 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526257 4745 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526271 4745 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526277 4745 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526285 4745 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526292 4745 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526298 4745 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.526304 4745 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526414 4745 flags.go:64] FLAG: --address="0.0.0.0" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526427 4745 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526437 4745 flags.go:64] FLAG: --anonymous-auth="true" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526445 4745 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526454 4745 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526461 4745 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526470 4745 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526481 4745 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526487 4745 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526494 4745 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526500 4745 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526507 4745 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526514 4745 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526520 4745 flags.go:64] FLAG: --cgroup-root="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526525 4745 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526533 4745 flags.go:64] FLAG: --client-ca-file="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526539 4745 flags.go:64] FLAG: --cloud-config="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526545 4745 flags.go:64] FLAG: --cloud-provider="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526551 4745 flags.go:64] FLAG: --cluster-dns="[]" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526560 4745 flags.go:64] FLAG: --cluster-domain="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526566 4745 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526572 4745 flags.go:64] FLAG: --config-dir="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526579 4745 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526586 4745 flags.go:64] FLAG: --container-log-max-files="5" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526594 4745 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526600 4745 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526607 4745 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526614 4745 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526621 4745 flags.go:64] FLAG: --contention-profiling="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526627 4745 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526633 4745 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526640 4745 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526646 4745 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526653 4745 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526660 4745 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526666 4745 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526672 4745 flags.go:64] FLAG: --enable-load-reader="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526679 4745 flags.go:64] FLAG: --enable-server="true" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526685 4745 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526693 4745 flags.go:64] FLAG: --event-burst="100" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526699 4745 flags.go:64] FLAG: --event-qps="50" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526706 4745 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526712 4745 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526718 4745 flags.go:64] FLAG: --eviction-hard="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526725 4745 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526753 4745 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526759 4745 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526766 4745 flags.go:64] FLAG: --eviction-soft="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526773 4745 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526779 4745 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526785 4745 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526791 4745 flags.go:64] FLAG: --experimental-mounter-path="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526797 4745 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526803 4745 flags.go:64] FLAG: --fail-swap-on="true" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526810 4745 flags.go:64] FLAG: --feature-gates="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526818 4745 flags.go:64] FLAG: --file-check-frequency="20s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526824 4745 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526831 4745 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526837 4745 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526843 4745 flags.go:64] FLAG: --healthz-port="10248" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526850 4745 flags.go:64] FLAG: --help="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526856 4745 flags.go:64] FLAG: --hostname-override="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526862 4745 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526868 4745 flags.go:64] FLAG: --http-check-frequency="20s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526874 4745 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526880 4745 flags.go:64] FLAG: --image-credential-provider-config="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526886 4745 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526892 4745 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526900 4745 flags.go:64] FLAG: --image-service-endpoint="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526907 4745 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526913 4745 flags.go:64] FLAG: --kube-api-burst="100" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526920 4745 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526926 4745 flags.go:64] FLAG: --kube-api-qps="50" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526932 4745 flags.go:64] FLAG: --kube-reserved="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526938 4745 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526944 4745 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526950 4745 flags.go:64] FLAG: --kubelet-cgroups="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526956 4745 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526962 4745 flags.go:64] FLAG: --lock-file="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526968 4745 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526974 4745 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526981 4745 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.526997 4745 flags.go:64] FLAG: --log-json-split-stream="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527006 4745 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527013 4745 flags.go:64] FLAG: --log-text-split-stream="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527019 4745 flags.go:64] FLAG: --logging-format="text" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527031 4745 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527038 4745 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527045 4745 flags.go:64] FLAG: --manifest-url="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527051 4745 flags.go:64] FLAG: --manifest-url-header="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527060 4745 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527067 4745 flags.go:64] FLAG: --max-open-files="1000000" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527074 4745 flags.go:64] FLAG: --max-pods="110" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527080 4745 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527087 4745 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527093 4745 flags.go:64] FLAG: --memory-manager-policy="None" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527099 4745 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527106 4745 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527112 4745 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527118 4745 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527133 4745 flags.go:64] FLAG: --node-status-max-images="50" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527139 4745 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527146 4745 flags.go:64] FLAG: --oom-score-adj="-999" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527152 4745 flags.go:64] FLAG: --pod-cidr="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527158 4745 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527168 4745 flags.go:64] FLAG: --pod-manifest-path="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527174 4745 flags.go:64] FLAG: --pod-max-pids="-1" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527180 4745 flags.go:64] FLAG: --pods-per-core="0" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527186 4745 flags.go:64] FLAG: --port="10250" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527192 4745 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527199 4745 flags.go:64] FLAG: --provider-id="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527204 4745 flags.go:64] FLAG: --qos-reserved="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527210 4745 flags.go:64] FLAG: --read-only-port="10255" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527216 4745 flags.go:64] FLAG: --register-node="true" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527222 4745 flags.go:64] FLAG: --register-schedulable="true" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527229 4745 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527239 4745 flags.go:64] FLAG: --registry-burst="10" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527245 4745 flags.go:64] FLAG: --registry-qps="5" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527253 4745 flags.go:64] FLAG: --reserved-cpus="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527260 4745 flags.go:64] FLAG: --reserved-memory="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527268 4745 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527274 4745 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527280 4745 flags.go:64] FLAG: --rotate-certificates="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527286 4745 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527292 4745 flags.go:64] FLAG: --runonce="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527298 4745 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527305 4745 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527311 4745 flags.go:64] FLAG: --seccomp-default="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527317 4745 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527323 4745 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527330 4745 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527336 4745 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527343 4745 flags.go:64] FLAG: --storage-driver-password="root" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527349 4745 flags.go:64] FLAG: --storage-driver-secure="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527355 4745 flags.go:64] FLAG: --storage-driver-table="stats" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527361 4745 flags.go:64] FLAG: --storage-driver-user="root" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527367 4745 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527373 4745 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527379 4745 flags.go:64] FLAG: --system-cgroups="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527385 4745 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527395 4745 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527401 4745 flags.go:64] FLAG: --tls-cert-file="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527407 4745 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527415 4745 flags.go:64] FLAG: --tls-min-version="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527421 4745 flags.go:64] FLAG: --tls-private-key-file="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527427 4745 flags.go:64] FLAG: --topology-manager-policy="none" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527432 4745 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527439 4745 flags.go:64] FLAG: --topology-manager-scope="container" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527445 4745 flags.go:64] FLAG: --v="2" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527453 4745 flags.go:64] FLAG: --version="false" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527463 4745 flags.go:64] FLAG: --vmodule="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527470 4745 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.527477 4745 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527657 4745 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527667 4745 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527674 4745 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527681 4745 feature_gate.go:330] unrecognized feature gate: Example Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527687 4745 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527692 4745 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527698 4745 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527703 4745 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527708 4745 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527714 4745 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527719 4745 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527725 4745 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527754 4745 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527760 4745 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527766 4745 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527771 4745 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527776 4745 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527782 4745 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527787 4745 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527792 4745 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527798 4745 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527803 4745 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527808 4745 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527814 4745 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527819 4745 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527825 4745 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527831 4745 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527836 4745 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527842 4745 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527850 4745 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527856 4745 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527863 4745 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527869 4745 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527875 4745 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527881 4745 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527886 4745 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527891 4745 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527897 4745 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527903 4745 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527909 4745 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527914 4745 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527919 4745 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527924 4745 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527929 4745 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527937 4745 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527946 4745 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527953 4745 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527958 4745 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527965 4745 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527972 4745 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527978 4745 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527983 4745 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527989 4745 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.527994 4745 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.528000 4745 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.528005 4745 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.528011 4745 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.528016 4745 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.528023 4745 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.528030 4745 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.528035 4745 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.528043 4745 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.528049 4745 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.528055 4745 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.528060 4745 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.528066 4745 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.528071 4745 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.528077 4745 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.528082 4745 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.528087 4745 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.528092 4745 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.528102 4745 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.537922 4745 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.537957 4745 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538097 4745 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538111 4745 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538125 4745 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538134 4745 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538144 4745 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538153 4745 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538162 4745 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538170 4745 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538178 4745 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538186 4745 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538197 4745 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538207 4745 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538217 4745 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538226 4745 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538234 4745 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538242 4745 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538250 4745 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538258 4745 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538266 4745 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538274 4745 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538282 4745 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538290 4745 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538298 4745 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538306 4745 feature_gate.go:330] unrecognized feature gate: Example Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538314 4745 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538322 4745 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538330 4745 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538337 4745 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538345 4745 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538352 4745 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538360 4745 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538370 4745 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538377 4745 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538385 4745 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538394 4745 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538402 4745 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538410 4745 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538418 4745 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538425 4745 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538433 4745 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538440 4745 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538448 4745 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538459 4745 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538468 4745 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538476 4745 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538486 4745 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538495 4745 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538503 4745 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538511 4745 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538519 4745 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538527 4745 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538534 4745 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538543 4745 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538551 4745 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538559 4745 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538566 4745 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538574 4745 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538582 4745 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538590 4745 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538597 4745 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538605 4745 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538615 4745 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538625 4745 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538635 4745 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538643 4745 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538651 4745 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538659 4745 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538666 4745 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538675 4745 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538683 4745 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538692 4745 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.538705 4745 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538957 4745 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538968 4745 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538977 4745 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538986 4745 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.538994 4745 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539002 4745 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539011 4745 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539020 4745 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539028 4745 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539038 4745 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539048 4745 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539057 4745 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539068 4745 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539076 4745 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539084 4745 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539092 4745 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539100 4745 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539108 4745 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539116 4745 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539123 4745 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539131 4745 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539139 4745 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539147 4745 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539155 4745 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539164 4745 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539171 4745 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539180 4745 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539188 4745 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539196 4745 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539204 4745 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539212 4745 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539219 4745 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539227 4745 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539235 4745 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539244 4745 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539251 4745 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539259 4745 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539267 4745 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539275 4745 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539285 4745 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539295 4745 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539305 4745 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539314 4745 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539323 4745 feature_gate.go:330] unrecognized feature gate: Example Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539332 4745 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539341 4745 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539349 4745 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539358 4745 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539366 4745 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539373 4745 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539381 4745 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539388 4745 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539396 4745 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539404 4745 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539411 4745 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539419 4745 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539427 4745 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539435 4745 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539443 4745 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539450 4745 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539457 4745 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539465 4745 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539473 4745 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539481 4745 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539488 4745 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539496 4745 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539503 4745 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539511 4745 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539519 4745 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539526 4745 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.539535 4745 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.539546 4745 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.539778 4745 server.go:940] "Client rotation is on, will bootstrap in background" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.546338 4745 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.546449 4745 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.548449 4745 server.go:997] "Starting client certificate rotation" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.548486 4745 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.549525 4745 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-28 14:32:42.732684429 +0000 UTC Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.549659 4745 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1177h14m36.183031684s for next certificate rotation Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.578451 4745 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.581323 4745 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.597581 4745 log.go:25] "Validated CRI v1 runtime API" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.629578 4745 log.go:25] "Validated CRI v1 image API" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.631720 4745 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.637331 4745 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-10-13-13-55-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.637391 4745 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.656970 4745 manager.go:217] Machine: {Timestamp:2025-10-10 13:18:06.653328083 +0000 UTC m=+0.550984876 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:8fd7c1b2-971d-45b1-b219-6a04ef38b544 BootID:53df7a22-5107-4004-bd75-de252c4adb13 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:7a:d2:5f Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:7a:d2:5f Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:73:d4:93 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:e4:a9:c4 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:ea:5e:e1 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:31:bf:16 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:da:11:5c:d5:91:b4 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:22:fd:7d:96:61:29 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.657268 4745 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.657495 4745 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.659781 4745 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.659994 4745 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.660042 4745 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.660327 4745 topology_manager.go:138] "Creating topology manager with none policy" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.660342 4745 container_manager_linux.go:303] "Creating device plugin manager" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.660969 4745 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.661036 4745 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.661341 4745 state_mem.go:36] "Initialized new in-memory state store" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.661459 4745 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.664796 4745 kubelet.go:418] "Attempting to sync node with API server" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.664829 4745 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.664862 4745 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.664880 4745 kubelet.go:324] "Adding apiserver pod source" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.664896 4745 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.669040 4745 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.670540 4745 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.673271 4745 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.673777 4745 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.673797 4745 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 10 13:18:06 crc kubenswrapper[4745]: E1010 13:18:06.673897 4745 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 10 13:18:06 crc kubenswrapper[4745]: E1010 13:18:06.673934 4745 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.674796 4745 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.674915 4745 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.674993 4745 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.675073 4745 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.675147 4745 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.675211 4745 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.675273 4745 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.675332 4745 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.675386 4745 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.675440 4745 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.675492 4745 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.675540 4745 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.677631 4745 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.678408 4745 server.go:1280] "Started kubelet" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.679521 4745 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.679552 4745 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.679905 4745 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.680583 4745 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 10 13:18:06 crc systemd[1]: Started Kubernetes Kubelet. Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.682185 4745 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.682236 4745 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.682263 4745 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 18:20:20.945962915 +0000 UTC Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.682328 4745 server.go:460] "Adding debug handlers to kubelet server" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.682330 4745 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1781h2m14.26363463s for next certificate rotation Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.687690 4745 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.687723 4745 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.687966 4745 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 10 13:18:06 crc kubenswrapper[4745]: E1010 13:18:06.688064 4745 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="200ms" Oct 10 13:18:06 crc kubenswrapper[4745]: E1010 13:18:06.690169 4745 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.691761 4745 factory.go:55] Registering systemd factory Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.691786 4745 factory.go:221] Registration of the systemd container factory successfully Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.691741 4745 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 10 13:18:06 crc kubenswrapper[4745]: E1010 13:18:06.692112 4745 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 10 13:18:06 crc kubenswrapper[4745]: E1010 13:18:06.690078 4745 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.200:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186d23b44dd0b5fd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-10 13:18:06.678341117 +0000 UTC m=+0.575997880,LastTimestamp:2025-10-10 13:18:06.678341117 +0000 UTC m=+0.575997880,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.693618 4745 factory.go:153] Registering CRI-O factory Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.693675 4745 factory.go:221] Registration of the crio container factory successfully Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.693934 4745 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.694001 4745 factory.go:103] Registering Raw factory Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.694048 4745 manager.go:1196] Started watching for new ooms in manager Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.695402 4745 manager.go:319] Starting recovery of all containers Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.699823 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.699885 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.699905 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.699921 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.699937 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.699955 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.699971 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.699988 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700005 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700021 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700037 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700054 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700071 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700113 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700129 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700145 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700186 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700203 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700220 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700238 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700255 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700283 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700301 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700321 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700393 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700413 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700435 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700455 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700472 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700488 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700504 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700522 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700541 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700564 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700585 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700612 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700629 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700649 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700668 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700687 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700706 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700722 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700800 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700822 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700840 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.700861 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702462 4745 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702508 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702534 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702554 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702574 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702593 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702611 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702640 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702662 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702693 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702714 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702757 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702778 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702798 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702818 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702836 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702857 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702875 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702901 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702924 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702952 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.702978 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703009 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703034 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703080 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703101 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703121 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703142 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703162 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703191 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703212 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703233 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703253 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703273 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703304 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703323 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703341 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703360 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703377 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703405 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703422 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703441 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703465 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703484 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703511 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703530 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703548 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703566 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703583 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703610 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703645 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703849 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703875 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703894 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703918 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.703936 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704001 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704024 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704043 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704071 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704106 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704127 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704155 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704177 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704198 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704225 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704250 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704271 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704290 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704308 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704326 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704350 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704369 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704395 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704414 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704459 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704483 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704500 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704515 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704540 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704564 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704583 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704610 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704630 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704647 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704664 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704685 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704703 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704721 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704761 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704778 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704797 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704813 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704830 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704853 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704879 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704902 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704917 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704935 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704952 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704969 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.704987 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705002 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705020 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705067 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705087 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705105 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705124 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705141 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705158 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705178 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705197 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705215 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705233 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705253 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705270 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705287 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705303 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705319 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705337 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705354 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705371 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705392 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705411 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705431 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705449 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705468 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705486 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705511 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705530 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705550 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705571 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705587 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705604 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705620 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705638 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705656 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705705 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705724 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705802 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705823 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705842 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705860 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705878 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705898 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705915 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705946 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705964 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705981 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.705998 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.706019 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.706044 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.706064 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.706083 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.706100 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.706119 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.706138 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.706156 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.706173 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.706190 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.706209 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.706228 4745 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.706246 4745 reconstruct.go:97] "Volume reconstruction finished" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.706258 4745 reconciler.go:26] "Reconciler: start to sync state" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.725878 4745 manager.go:324] Recovery completed Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.741040 4745 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.741315 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.743593 4745 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.743657 4745 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.743695 4745 kubelet.go:2335] "Starting kubelet main sync loop" Oct 10 13:18:06 crc kubenswrapper[4745]: E1010 13:18:06.743823 4745 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.746642 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.746678 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.746699 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:06 crc kubenswrapper[4745]: W1010 13:18:06.748534 4745 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.748631 4745 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.748675 4745 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.748710 4745 state_mem.go:36] "Initialized new in-memory state store" Oct 10 13:18:06 crc kubenswrapper[4745]: E1010 13:18:06.748645 4745 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.774989 4745 policy_none.go:49] "None policy: Start" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.776514 4745 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.776545 4745 state_mem.go:35] "Initializing new in-memory state store" Oct 10 13:18:06 crc kubenswrapper[4745]: E1010 13:18:06.790912 4745 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.834664 4745 manager.go:334] "Starting Device Plugin manager" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.834761 4745 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.834784 4745 server.go:79] "Starting device plugin registration server" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.835509 4745 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.835535 4745 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.836878 4745 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.837086 4745 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.837101 4745 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.844663 4745 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.844766 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.845850 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.845888 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.845901 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.846117 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.846500 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.846623 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:06 crc kubenswrapper[4745]: E1010 13:18:06.846787 4745 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.847183 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.847206 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.847217 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.847335 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.847612 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.847696 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.848603 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.848635 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.848652 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.848814 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.848936 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.848976 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.849624 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.849665 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.849681 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.849701 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.849687 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.849712 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.849758 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.849798 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.849815 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.850143 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.850295 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.850342 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.850705 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.850776 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.850795 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.850950 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.850984 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.851002 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.851049 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.851071 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.851083 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.851335 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.851377 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.852153 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.852185 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.852200 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:06 crc kubenswrapper[4745]: E1010 13:18:06.889331 4745 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="400ms" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.908050 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.908117 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.908155 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.908193 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.908274 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.908336 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.908429 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.908456 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.908514 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.908641 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.908766 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.908834 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.908901 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.908953 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.909008 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.936196 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.937587 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.937636 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.937650 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:06 crc kubenswrapper[4745]: I1010 13:18:06.937689 4745 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 13:18:06 crc kubenswrapper[4745]: E1010 13:18:06.938460 4745 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.200:6443: connect: connection refused" node="crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010492 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010564 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010599 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010622 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010642 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010662 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010681 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010697 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010703 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010806 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010721 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010840 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010880 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010905 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010908 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010818 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010940 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010911 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010914 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.010950 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.011047 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.011067 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.011115 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.011144 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.011179 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.011156 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.011248 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.011316 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.011357 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.011441 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.139294 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.141067 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.141117 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.141131 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.141163 4745 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 13:18:07 crc kubenswrapper[4745]: E1010 13:18:07.142002 4745 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.200:6443: connect: connection refused" node="crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.186802 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.217824 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.228299 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: W1010 13:18:07.235951 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-1462499efa176eb862f3b3dfc369862ad2225103a54dd667dad26b36a9267092 WatchSource:0}: Error finding container 1462499efa176eb862f3b3dfc369862ad2225103a54dd667dad26b36a9267092: Status 404 returned error can't find the container with id 1462499efa176eb862f3b3dfc369862ad2225103a54dd667dad26b36a9267092 Oct 10 13:18:07 crc kubenswrapper[4745]: W1010 13:18:07.251342 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-34cfbfb5649b04e05480ba3d90f68543d83dec681fee09b6d83ed45bbd176295 WatchSource:0}: Error finding container 34cfbfb5649b04e05480ba3d90f68543d83dec681fee09b6d83ed45bbd176295: Status 404 returned error can't find the container with id 34cfbfb5649b04e05480ba3d90f68543d83dec681fee09b6d83ed45bbd176295 Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.254200 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.263613 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:07 crc kubenswrapper[4745]: W1010 13:18:07.279766 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-be7e7d40ed40b8b09e5c1a16ad7461032c96e6e11ade6e435f8c8ef5cde0a7bf WatchSource:0}: Error finding container be7e7d40ed40b8b09e5c1a16ad7461032c96e6e11ade6e435f8c8ef5cde0a7bf: Status 404 returned error can't find the container with id be7e7d40ed40b8b09e5c1a16ad7461032c96e6e11ade6e435f8c8ef5cde0a7bf Oct 10 13:18:07 crc kubenswrapper[4745]: E1010 13:18:07.290513 4745 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="800ms" Oct 10 13:18:07 crc kubenswrapper[4745]: W1010 13:18:07.294172 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-596c60cca829f56b83ff85e6bab621d3f81f703a498c3d841967f031c4067973 WatchSource:0}: Error finding container 596c60cca829f56b83ff85e6bab621d3f81f703a498c3d841967f031c4067973: Status 404 returned error can't find the container with id 596c60cca829f56b83ff85e6bab621d3f81f703a498c3d841967f031c4067973 Oct 10 13:18:07 crc kubenswrapper[4745]: W1010 13:18:07.494621 4745 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 10 13:18:07 crc kubenswrapper[4745]: E1010 13:18:07.494720 4745 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.542800 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.546845 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.546893 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.546912 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.546950 4745 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 13:18:07 crc kubenswrapper[4745]: E1010 13:18:07.547562 4745 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.200:6443: connect: connection refused" node="crc" Oct 10 13:18:07 crc kubenswrapper[4745]: W1010 13:18:07.558551 4745 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 10 13:18:07 crc kubenswrapper[4745]: E1010 13:18:07.558653 4745 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.681509 4745 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.756418 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"1b893f38eaf8e46eb75c04b4c951eb601c02919c54171552cb5cf72e62455697"} Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.758338 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"34cfbfb5649b04e05480ba3d90f68543d83dec681fee09b6d83ed45bbd176295"} Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.760199 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1462499efa176eb862f3b3dfc369862ad2225103a54dd667dad26b36a9267092"} Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.761599 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"596c60cca829f56b83ff85e6bab621d3f81f703a498c3d841967f031c4067973"} Oct 10 13:18:07 crc kubenswrapper[4745]: I1010 13:18:07.763206 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"be7e7d40ed40b8b09e5c1a16ad7461032c96e6e11ade6e435f8c8ef5cde0a7bf"} Oct 10 13:18:07 crc kubenswrapper[4745]: W1010 13:18:07.803590 4745 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 10 13:18:07 crc kubenswrapper[4745]: E1010 13:18:07.803672 4745 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 10 13:18:07 crc kubenswrapper[4745]: W1010 13:18:07.909921 4745 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 10 13:18:07 crc kubenswrapper[4745]: E1010 13:18:07.910440 4745 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 10 13:18:08 crc kubenswrapper[4745]: E1010 13:18:08.092387 4745 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="1.6s" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.347721 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.349397 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.349450 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.349471 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.349505 4745 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 13:18:08 crc kubenswrapper[4745]: E1010 13:18:08.350071 4745 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.200:6443: connect: connection refused" node="crc" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.680620 4745 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.771029 4745 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="701af5e4ef1c9e8c4f58ef5222fb6aee23624a16575e5b307ec14672c4143bff" exitCode=0 Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.771141 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"701af5e4ef1c9e8c4f58ef5222fb6aee23624a16575e5b307ec14672c4143bff"} Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.771246 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.772545 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.772585 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.772596 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.776384 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08"} Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.776428 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc"} Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.776442 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771"} Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.776454 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b"} Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.776543 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.778140 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.778199 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.778211 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.779703 4745 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407" exitCode=0 Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.780010 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407"} Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.780564 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.783058 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.783101 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.783119 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.785204 4745 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa" exitCode=0 Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.785333 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.785312 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa"} Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.786752 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.786791 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.786801 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.791670 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.794822 4745 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8" exitCode=0 Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.794962 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.795025 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8"} Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.796306 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.796366 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.796394 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.796784 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.796809 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:08 crc kubenswrapper[4745]: I1010 13:18:08.796819 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:09 crc kubenswrapper[4745]: W1010 13:18:09.258636 4745 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 10 13:18:09 crc kubenswrapper[4745]: E1010 13:18:09.258719 4745 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 10 13:18:09 crc kubenswrapper[4745]: W1010 13:18:09.565155 4745 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 10 13:18:09 crc kubenswrapper[4745]: E1010 13:18:09.565339 4745 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.680861 4745 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 10 13:18:09 crc kubenswrapper[4745]: E1010 13:18:09.696172 4745 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="3.2s" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.748713 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.801337 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4"} Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.801417 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7"} Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.801439 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f"} Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.801460 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd"} Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.804424 4745 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579" exitCode=0 Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.804522 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579"} Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.804595 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.806001 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.806045 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.806062 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.807122 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"81717c65162b43bab3af2ecbae6ec4e49ab3e89c9631187731fe36770d4c97c2"} Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.807184 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.813484 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.813522 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.813534 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.817090 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"57f6659c3fefc2b4892b34a407a26418aa86c780a7e03c9120794f4396c08913"} Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.817166 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"eab4465f351419e32376cc0b2062225c6faac1e27dce51bcf3c1ca4bb590fc4e"} Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.817185 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"193833f7e032a32aae576702c6965f7ea2869114b86d1a0349e7ad787869c74d"} Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.817111 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.817249 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.819071 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.819132 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.819157 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.821015 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.821048 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.821058 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.950215 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.951533 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.951559 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.951568 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:09 crc kubenswrapper[4745]: I1010 13:18:09.951712 4745 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 13:18:09 crc kubenswrapper[4745]: E1010 13:18:09.952283 4745 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.200:6443: connect: connection refused" node="crc" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.825332 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093"} Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.825425 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.826714 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.826816 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.826855 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.829092 4745 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9" exitCode=0 Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.829226 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.829255 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.829349 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.829506 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9"} Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.829627 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.830110 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.830788 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.830838 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.830855 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.830796 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.830986 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.830998 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.831035 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.831060 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.831007 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.831646 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.831699 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:10 crc kubenswrapper[4745]: I1010 13:18:10.831765 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:11 crc kubenswrapper[4745]: I1010 13:18:11.840627 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea"} Oct 10 13:18:11 crc kubenswrapper[4745]: I1010 13:18:11.840714 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd"} Oct 10 13:18:11 crc kubenswrapper[4745]: I1010 13:18:11.840760 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539"} Oct 10 13:18:11 crc kubenswrapper[4745]: I1010 13:18:11.840774 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89"} Oct 10 13:18:11 crc kubenswrapper[4745]: I1010 13:18:11.840806 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:11 crc kubenswrapper[4745]: I1010 13:18:11.840898 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:11 crc kubenswrapper[4745]: I1010 13:18:11.840924 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:11 crc kubenswrapper[4745]: I1010 13:18:11.842054 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:11 crc kubenswrapper[4745]: I1010 13:18:11.842132 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:11 crc kubenswrapper[4745]: I1010 13:18:11.842159 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:11 crc kubenswrapper[4745]: I1010 13:18:11.842536 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:11 crc kubenswrapper[4745]: I1010 13:18:11.842593 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:11 crc kubenswrapper[4745]: I1010 13:18:11.842614 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:12 crc kubenswrapper[4745]: I1010 13:18:12.694821 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:12 crc kubenswrapper[4745]: I1010 13:18:12.749148 4745 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 10 13:18:12 crc kubenswrapper[4745]: I1010 13:18:12.749312 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 13:18:12 crc kubenswrapper[4745]: I1010 13:18:12.851133 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a"} Oct 10 13:18:12 crc kubenswrapper[4745]: I1010 13:18:12.851232 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:12 crc kubenswrapper[4745]: I1010 13:18:12.851369 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:12 crc kubenswrapper[4745]: I1010 13:18:12.852781 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:12 crc kubenswrapper[4745]: I1010 13:18:12.852858 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:12 crc kubenswrapper[4745]: I1010 13:18:12.852883 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:12 crc kubenswrapper[4745]: I1010 13:18:12.853317 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:12 crc kubenswrapper[4745]: I1010 13:18:12.853363 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:12 crc kubenswrapper[4745]: I1010 13:18:12.853386 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:13 crc kubenswrapper[4745]: I1010 13:18:13.152471 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:13 crc kubenswrapper[4745]: I1010 13:18:13.154663 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:13 crc kubenswrapper[4745]: I1010 13:18:13.154725 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:13 crc kubenswrapper[4745]: I1010 13:18:13.154781 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:13 crc kubenswrapper[4745]: I1010 13:18:13.154823 4745 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 13:18:13 crc kubenswrapper[4745]: I1010 13:18:13.435793 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:13 crc kubenswrapper[4745]: I1010 13:18:13.855163 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:13 crc kubenswrapper[4745]: I1010 13:18:13.855164 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:13 crc kubenswrapper[4745]: I1010 13:18:13.857209 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:13 crc kubenswrapper[4745]: I1010 13:18:13.857283 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:13 crc kubenswrapper[4745]: I1010 13:18:13.857317 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:13 crc kubenswrapper[4745]: I1010 13:18:13.857780 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:13 crc kubenswrapper[4745]: I1010 13:18:13.857847 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:13 crc kubenswrapper[4745]: I1010 13:18:13.857867 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:14 crc kubenswrapper[4745]: I1010 13:18:14.442855 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 13:18:14 crc kubenswrapper[4745]: I1010 13:18:14.443041 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:14 crc kubenswrapper[4745]: I1010 13:18:14.444716 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:14 crc kubenswrapper[4745]: I1010 13:18:14.444823 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:14 crc kubenswrapper[4745]: I1010 13:18:14.444844 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:14 crc kubenswrapper[4745]: I1010 13:18:14.748799 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 10 13:18:14 crc kubenswrapper[4745]: I1010 13:18:14.858306 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:14 crc kubenswrapper[4745]: I1010 13:18:14.858372 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:14 crc kubenswrapper[4745]: I1010 13:18:14.860351 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:14 crc kubenswrapper[4745]: I1010 13:18:14.860402 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:14 crc kubenswrapper[4745]: I1010 13:18:14.860422 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:14 crc kubenswrapper[4745]: I1010 13:18:14.861226 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:14 crc kubenswrapper[4745]: I1010 13:18:14.861308 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:14 crc kubenswrapper[4745]: I1010 13:18:14.861330 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:16 crc kubenswrapper[4745]: I1010 13:18:16.148545 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 13:18:16 crc kubenswrapper[4745]: I1010 13:18:16.148933 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:16 crc kubenswrapper[4745]: I1010 13:18:16.150600 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:16 crc kubenswrapper[4745]: I1010 13:18:16.150674 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:16 crc kubenswrapper[4745]: I1010 13:18:16.150690 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:16 crc kubenswrapper[4745]: I1010 13:18:16.157300 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 13:18:16 crc kubenswrapper[4745]: I1010 13:18:16.590525 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 13:18:16 crc kubenswrapper[4745]: E1010 13:18:16.846953 4745 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 10 13:18:16 crc kubenswrapper[4745]: I1010 13:18:16.864180 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:16 crc kubenswrapper[4745]: I1010 13:18:16.864971 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:16 crc kubenswrapper[4745]: I1010 13:18:16.865011 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:16 crc kubenswrapper[4745]: I1010 13:18:16.865022 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:17 crc kubenswrapper[4745]: I1010 13:18:17.553482 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 10 13:18:17 crc kubenswrapper[4745]: I1010 13:18:17.553904 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:17 crc kubenswrapper[4745]: I1010 13:18:17.556163 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:17 crc kubenswrapper[4745]: I1010 13:18:17.556243 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:17 crc kubenswrapper[4745]: I1010 13:18:17.556274 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:17 crc kubenswrapper[4745]: I1010 13:18:17.866402 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:17 crc kubenswrapper[4745]: I1010 13:18:17.868803 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:17 crc kubenswrapper[4745]: I1010 13:18:17.868869 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:17 crc kubenswrapper[4745]: I1010 13:18:17.868893 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:17 crc kubenswrapper[4745]: I1010 13:18:17.873223 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 13:18:18 crc kubenswrapper[4745]: I1010 13:18:18.869398 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:18 crc kubenswrapper[4745]: I1010 13:18:18.870239 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:18 crc kubenswrapper[4745]: I1010 13:18:18.870269 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:18 crc kubenswrapper[4745]: I1010 13:18:18.870280 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:20 crc kubenswrapper[4745]: W1010 13:18:20.249765 4745 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 10 13:18:20 crc kubenswrapper[4745]: I1010 13:18:20.249921 4745 trace.go:236] Trace[1760303880]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Oct-2025 13:18:10.248) (total time: 10001ms): Oct 10 13:18:20 crc kubenswrapper[4745]: Trace[1760303880]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (13:18:20.249) Oct 10 13:18:20 crc kubenswrapper[4745]: Trace[1760303880]: [10.001529582s] [10.001529582s] END Oct 10 13:18:20 crc kubenswrapper[4745]: E1010 13:18:20.249960 4745 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 10 13:18:20 crc kubenswrapper[4745]: E1010 13:18:20.319145 4745 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.186d23b44dd0b5fd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-10 13:18:06.678341117 +0000 UTC m=+0.575997880,LastTimestamp:2025-10-10 13:18:06.678341117 +0000 UTC m=+0.575997880,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 10 13:18:20 crc kubenswrapper[4745]: W1010 13:18:20.347898 4745 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 10 13:18:20 crc kubenswrapper[4745]: I1010 13:18:20.348063 4745 trace.go:236] Trace[55394459]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Oct-2025 13:18:10.346) (total time: 10001ms): Oct 10 13:18:20 crc kubenswrapper[4745]: Trace[55394459]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (13:18:20.347) Oct 10 13:18:20 crc kubenswrapper[4745]: Trace[55394459]: [10.001886111s] [10.001886111s] END Oct 10 13:18:20 crc kubenswrapper[4745]: E1010 13:18:20.348097 4745 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 10 13:18:20 crc kubenswrapper[4745]: I1010 13:18:20.681923 4745 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 10 13:18:21 crc kubenswrapper[4745]: I1010 13:18:21.208862 4745 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 10 13:18:21 crc kubenswrapper[4745]: I1010 13:18:21.208938 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 10 13:18:21 crc kubenswrapper[4745]: I1010 13:18:21.213593 4745 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 10 13:18:21 crc kubenswrapper[4745]: I1010 13:18:21.213669 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 10 13:18:22 crc kubenswrapper[4745]: I1010 13:18:22.712206 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:22 crc kubenswrapper[4745]: I1010 13:18:22.712436 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:22 crc kubenswrapper[4745]: I1010 13:18:22.714080 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:22 crc kubenswrapper[4745]: I1010 13:18:22.714181 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:22 crc kubenswrapper[4745]: I1010 13:18:22.714204 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:22 crc kubenswrapper[4745]: I1010 13:18:22.720117 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:22 crc kubenswrapper[4745]: I1010 13:18:22.750064 4745 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 10 13:18:22 crc kubenswrapper[4745]: I1010 13:18:22.750159 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 10 13:18:22 crc kubenswrapper[4745]: I1010 13:18:22.881066 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:22 crc kubenswrapper[4745]: I1010 13:18:22.882486 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:22 crc kubenswrapper[4745]: I1010 13:18:22.882542 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:22 crc kubenswrapper[4745]: I1010 13:18:22.882565 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:23 crc kubenswrapper[4745]: I1010 13:18:23.980917 4745 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.318586 4745 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.674687 4745 apiserver.go:52] "Watching apiserver" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.684698 4745 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.685137 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.685661 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.685759 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.685852 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:24 crc kubenswrapper[4745]: E1010 13:18:24.685887 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:24 crc kubenswrapper[4745]: E1010 13:18:24.685952 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.685979 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.686076 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.686593 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:24 crc kubenswrapper[4745]: E1010 13:18:24.686659 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.689511 4745 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.689554 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.690048 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.690283 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.690309 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.690408 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.691265 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.691815 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.691871 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.692102 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.732484 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.755514 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.779261 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.787005 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.795838 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.808499 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.809140 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.812598 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.825931 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.845956 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.865143 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.881565 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.896216 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.913619 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.942063 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.960659 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.976664 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:24 crc kubenswrapper[4745]: I1010 13:18:24.993364 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:25 crc kubenswrapper[4745]: I1010 13:18:25.008105 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:25 crc kubenswrapper[4745]: I1010 13:18:25.744263 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:25 crc kubenswrapper[4745]: E1010 13:18:25.744449 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.210055 4745 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.211815 4745 trace.go:236] Trace[1867864649]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Oct-2025 13:18:14.765) (total time: 11445ms): Oct 10 13:18:26 crc kubenswrapper[4745]: Trace[1867864649]: ---"Objects listed" error: 11445ms (13:18:26.211) Oct 10 13:18:26 crc kubenswrapper[4745]: Trace[1867864649]: [11.445887628s] [11.445887628s] END Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.211855 4745 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.213360 4745 trace.go:236] Trace[1135705106]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Oct-2025 13:18:13.615) (total time: 12598ms): Oct 10 13:18:26 crc kubenswrapper[4745]: Trace[1135705106]: ---"Objects listed" error: 12598ms (13:18:26.213) Oct 10 13:18:26 crc kubenswrapper[4745]: Trace[1135705106]: [12.598168085s] [12.598168085s] END Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.213400 4745 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.214861 4745 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.215202 4745 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.244629 4745 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:34886->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.244772 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:34886->192.168.126.11:17697: read: connection reset by peer" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.245273 4745 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.245385 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.316706 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.316859 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.316897 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.316937 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.316976 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317007 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317108 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317146 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317180 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317216 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317218 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317247 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317362 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317431 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317457 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317484 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317506 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317532 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317556 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317582 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317604 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317625 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317645 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317649 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317667 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317715 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317767 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317864 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317916 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317913 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.317954 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318003 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318020 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318039 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318074 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318122 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318176 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318223 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318230 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318298 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318338 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318385 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318439 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318493 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318543 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318656 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318775 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318971 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319065 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319101 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319167 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319206 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319292 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319336 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319373 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319406 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319442 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319474 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319510 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319545 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319642 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319680 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319716 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319788 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319824 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319877 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319929 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319966 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319999 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320035 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320075 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320129 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320180 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320233 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318410 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318540 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320290 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318653 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318842 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.318852 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319007 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319102 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319115 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319314 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319374 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319485 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319518 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319524 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319557 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.319914 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320030 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320046 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320433 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320060 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320086 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320207 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320210 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320249 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.320295 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:18:26.820254681 +0000 UTC m=+20.717911614 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320548 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320581 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320605 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320634 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320663 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320685 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320703 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320713 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320815 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320851 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320889 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320891 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320925 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320965 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320999 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321029 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321044 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321081 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321151 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321187 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321223 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321258 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321297 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321332 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321364 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321399 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321434 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321466 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321502 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321535 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321572 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321608 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321643 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321679 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321717 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321827 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322105 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322145 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322183 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322217 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322252 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322289 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322322 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322356 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322393 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322428 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322461 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322496 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322538 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322575 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322611 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322649 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322685 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322723 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322816 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322855 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322892 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322929 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322964 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323001 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323037 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323072 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323107 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323148 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323185 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323218 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323255 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323315 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323350 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323385 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323419 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323452 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323488 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323522 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323561 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323596 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323633 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323666 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323704 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324036 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324079 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324113 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324146 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324184 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324224 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324259 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324325 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324368 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324404 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324439 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324474 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324514 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324833 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324880 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324918 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324955 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.324989 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325025 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325059 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325096 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325132 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325172 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325209 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325243 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325289 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325324 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325360 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325401 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325436 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325476 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325531 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325568 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325603 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325641 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325676 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325712 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325787 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325823 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325861 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325898 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325937 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325972 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326009 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326046 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326081 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326118 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326159 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326198 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326237 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326272 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326352 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326430 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326476 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326514 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326565 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326606 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326644 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326690 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326757 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326801 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326838 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326883 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326925 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326963 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327045 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327072 4745 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327098 4745 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327123 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327147 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327171 4745 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327195 4745 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327215 4745 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327237 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327257 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327278 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327298 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327321 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327345 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327366 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327387 4745 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327410 4745 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327430 4745 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327452 4745 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327473 4745 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327494 4745 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327515 4745 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327534 4745 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327556 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327580 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327601 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327621 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327641 4745 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327665 4745 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327688 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327708 4745 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327767 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327788 4745 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.320393 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.328001 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321464 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321537 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321702 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321779 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.321837 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322679 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322748 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.322723 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323050 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323061 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323096 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323450 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323530 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.323859 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325684 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.328347 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.328639 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327990 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.328881 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325834 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325933 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326126 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326193 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326326 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326651 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326820 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.326855 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327089 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327262 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327323 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327393 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327528 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327564 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327668 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327748 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327962 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.327968 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.328009 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.328904 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.325758 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.331316 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.331532 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.331724 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.331801 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.331862 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.332033 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.332058 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.332200 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.332364 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.332451 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.332578 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.332636 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.332782 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.332795 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.333137 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.333223 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.333358 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.333441 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.333580 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.333822 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.334155 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.334579 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.334622 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.335105 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.335101 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.335374 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.335517 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.335549 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.336574 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.336812 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.336910 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.337331 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.337359 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.337484 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.337602 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.337634 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.337753 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.337798 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.337821 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.337975 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.338390 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.338403 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.338518 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.338785 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.338835 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.339184 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.339378 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.339780 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.339855 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.339881 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.339838 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.340022 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.340062 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.340122 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.340366 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.340673 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.340870 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.341265 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.341428 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.341341 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.342019 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.342229 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.342479 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.342554 4745 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.343328 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.343341 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.343527 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.344015 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.344258 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.344055 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.345325 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.345578 4745 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.345610 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.345909 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.346144 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.346383 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.346547 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.346584 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:26.845637655 +0000 UTC m=+20.743294418 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.346663 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.346692 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.347116 4745 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.347274 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.347320 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:26.847305047 +0000 UTC m=+20.744961810 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.347441 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.347518 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.347571 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.347584 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.347761 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.347827 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.348202 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.348218 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.348282 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.348367 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.348456 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.348618 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.348720 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.349214 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.350091 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.350214 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.350508 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.350720 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.357602 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.358325 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.358562 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.358673 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.358991 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.359269 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.359561 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.359766 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.360391 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.360424 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.360446 4745 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.360468 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.360547 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:26.860509557 +0000 UTC m=+20.758166340 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.360704 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.360770 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.360609 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.360831 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.360852 4745 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.360916 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:26.860898296 +0000 UTC m=+20.758555069 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.360664 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.360956 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.360997 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.361163 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.361261 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.362041 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.363186 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.363261 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.363568 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.363977 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.364539 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.365447 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.366278 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.367171 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.371437 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.373811 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.374077 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.382973 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.394968 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.397286 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429267 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429326 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429382 4745 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429396 4745 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429448 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429505 4745 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429607 4745 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429635 4745 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429652 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429669 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429686 4745 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429703 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429710 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429720 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429786 4745 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429804 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429821 4745 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429837 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429850 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429862 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429874 4745 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429886 4745 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429898 4745 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429910 4745 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429922 4745 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429934 4745 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429947 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429959 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429971 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429983 4745 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.429996 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430012 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430025 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430041 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430054 4745 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430065 4745 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430077 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430090 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430102 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430114 4745 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430127 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430140 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430152 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430165 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430178 4745 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430190 4745 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430202 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430214 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430226 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430238 4745 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430250 4745 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430262 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430274 4745 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430287 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430299 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430311 4745 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430323 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430336 4745 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430347 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430359 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430371 4745 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430382 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430396 4745 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430409 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430421 4745 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430433 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430444 4745 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430456 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430467 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430479 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430490 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430502 4745 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430514 4745 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430526 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430537 4745 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430550 4745 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430563 4745 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430574 4745 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430585 4745 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430598 4745 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430609 4745 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430620 4745 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430632 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430644 4745 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430656 4745 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430668 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430679 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430693 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430704 4745 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430715 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430745 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430758 4745 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430769 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430780 4745 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430791 4745 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430802 4745 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430813 4745 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430839 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430850 4745 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430861 4745 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430872 4745 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430884 4745 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430895 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430906 4745 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430918 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430929 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430940 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430951 4745 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430962 4745 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430974 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430984 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.430996 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431008 4745 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431020 4745 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431032 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431044 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431057 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431068 4745 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431081 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431092 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431103 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431114 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431126 4745 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431137 4745 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431149 4745 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431160 4745 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431173 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431185 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431197 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431209 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431221 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431233 4745 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431251 4745 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431263 4745 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431274 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431285 4745 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431297 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431309 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431320 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431331 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431343 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431355 4745 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431367 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431378 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431391 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431405 4745 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431416 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431428 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431440 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431452 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431463 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431474 4745 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431485 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431497 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431509 4745 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431521 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431534 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431545 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431556 4745 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431567 4745 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431579 4745 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431590 4745 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431601 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431614 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.431626 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.509060 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 13:18:26 crc kubenswrapper[4745]: W1010 13:18:26.522425 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-2f21b4242177b0c34c632a5b0554301aa483dcc0c392354c3230876062d8aa72 WatchSource:0}: Error finding container 2f21b4242177b0c34c632a5b0554301aa483dcc0c392354c3230876062d8aa72: Status 404 returned error can't find the container with id 2f21b4242177b0c34c632a5b0554301aa483dcc0c392354c3230876062d8aa72 Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.526355 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.537548 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 13:18:26 crc kubenswrapper[4745]: W1010 13:18:26.538205 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-250c79ee54baf2937448b0481dcfcde72530e89104f7064fb9e1579fa94c8b7e WatchSource:0}: Error finding container 250c79ee54baf2937448b0481dcfcde72530e89104f7064fb9e1579fa94c8b7e: Status 404 returned error can't find the container with id 250c79ee54baf2937448b0481dcfcde72530e89104f7064fb9e1579fa94c8b7e Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.743925 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.743960 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.744371 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.744501 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.747939 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.748810 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.749495 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.750509 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.751109 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.752068 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.752632 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.753697 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.754545 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.755312 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.755480 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.756039 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.757485 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.758011 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.758549 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.759464 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.759984 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.760914 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.761297 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.761860 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.762837 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.763307 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.764060 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.764103 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.765215 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.766084 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.766953 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.767594 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.768605 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.769108 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.770095 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.770597 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.771077 4745 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.771518 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.773443 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.773975 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.775156 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.776868 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.776948 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.777757 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.778797 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.779532 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.781540 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.782489 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.783297 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.784064 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.785621 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.786766 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.787744 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.788247 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.789692 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.790372 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.791043 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.792101 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.792272 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.792815 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.793824 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.794295 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.802574 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.813309 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.822267 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.834839 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.835001 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:18:27.834985275 +0000 UTC m=+21.732642038 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.895969 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171"} Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.896031 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2f21b4242177b0c34c632a5b0554301aa483dcc0c392354c3230876062d8aa72"} Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.899344 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.902751 4745 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093" exitCode=255 Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.902828 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093"} Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.905026 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f"} Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.905069 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791"} Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.905100 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a51f52058e4d28e0027ccfd36cdf5529687801849fa889411baa762e873a30a6"} Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.905982 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"250c79ee54baf2937448b0481dcfcde72530e89104f7064fb9e1579fa94c8b7e"} Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.912913 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.913310 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.915535 4745 scope.go:117] "RemoveContainer" containerID="72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.926632 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.935391 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.935469 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.935517 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.935570 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.935690 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.935707 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.935717 4745 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.935787 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:27.935769904 +0000 UTC m=+21.833426667 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.935780 4745 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.935946 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:27.935883957 +0000 UTC m=+21.833540740 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.936129 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.936179 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.936199 4745 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.936271 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:27.936229136 +0000 UTC m=+21.833885929 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.936381 4745 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 13:18:26 crc kubenswrapper[4745]: E1010 13:18:26.936463 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:27.936448411 +0000 UTC m=+21.834105234 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.947805 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.972749 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:26 crc kubenswrapper[4745]: I1010 13:18:26.988539 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.002906 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.032817 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.045794 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.058031 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.073979 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.085061 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.095341 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.105125 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.114892 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.125111 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.358472 4745 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.398550 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-7rdsn"] Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.398937 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7rdsn" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.402357 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.402426 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.402507 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.409747 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.421769 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:27Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.434323 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:27Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.438822 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bce814bc-030f-402e-ae0a-546f1c63edbd-serviceca\") pod \"node-ca-7rdsn\" (UID: \"bce814bc-030f-402e-ae0a-546f1c63edbd\") " pod="openshift-image-registry/node-ca-7rdsn" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.438902 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bce814bc-030f-402e-ae0a-546f1c63edbd-host\") pod \"node-ca-7rdsn\" (UID: \"bce814bc-030f-402e-ae0a-546f1c63edbd\") " pod="openshift-image-registry/node-ca-7rdsn" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.438979 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9ltc\" (UniqueName: \"kubernetes.io/projected/bce814bc-030f-402e-ae0a-546f1c63edbd-kube-api-access-c9ltc\") pod \"node-ca-7rdsn\" (UID: \"bce814bc-030f-402e-ae0a-546f1c63edbd\") " pod="openshift-image-registry/node-ca-7rdsn" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.448226 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:27Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.461169 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:27Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.474247 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:27Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.498605 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:27Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.521126 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:27Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.534540 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:27Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.540022 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bce814bc-030f-402e-ae0a-546f1c63edbd-host\") pod \"node-ca-7rdsn\" (UID: \"bce814bc-030f-402e-ae0a-546f1c63edbd\") " pod="openshift-image-registry/node-ca-7rdsn" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.540065 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9ltc\" (UniqueName: \"kubernetes.io/projected/bce814bc-030f-402e-ae0a-546f1c63edbd-kube-api-access-c9ltc\") pod \"node-ca-7rdsn\" (UID: \"bce814bc-030f-402e-ae0a-546f1c63edbd\") " pod="openshift-image-registry/node-ca-7rdsn" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.540131 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bce814bc-030f-402e-ae0a-546f1c63edbd-host\") pod \"node-ca-7rdsn\" (UID: \"bce814bc-030f-402e-ae0a-546f1c63edbd\") " pod="openshift-image-registry/node-ca-7rdsn" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.540088 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bce814bc-030f-402e-ae0a-546f1c63edbd-serviceca\") pod \"node-ca-7rdsn\" (UID: \"bce814bc-030f-402e-ae0a-546f1c63edbd\") " pod="openshift-image-registry/node-ca-7rdsn" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.541283 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bce814bc-030f-402e-ae0a-546f1c63edbd-serviceca\") pod \"node-ca-7rdsn\" (UID: \"bce814bc-030f-402e-ae0a-546f1c63edbd\") " pod="openshift-image-registry/node-ca-7rdsn" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.550723 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:27Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.560361 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9ltc\" (UniqueName: \"kubernetes.io/projected/bce814bc-030f-402e-ae0a-546f1c63edbd-kube-api-access-c9ltc\") pod \"node-ca-7rdsn\" (UID: \"bce814bc-030f-402e-ae0a-546f1c63edbd\") " pod="openshift-image-registry/node-ca-7rdsn" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.713228 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7rdsn" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.744376 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:27 crc kubenswrapper[4745]: E1010 13:18:27.744513 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:27 crc kubenswrapper[4745]: W1010 13:18:27.814957 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbce814bc_030f_402e_ae0a_546f1c63edbd.slice/crio-53dab70a12bb25e3fa2611db860662c20975e597c052bc87ced71039bca9ff16 WatchSource:0}: Error finding container 53dab70a12bb25e3fa2611db860662c20975e597c052bc87ced71039bca9ff16: Status 404 returned error can't find the container with id 53dab70a12bb25e3fa2611db860662c20975e597c052bc87ced71039bca9ff16 Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.843474 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:18:27 crc kubenswrapper[4745]: E1010 13:18:27.843601 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:18:29.843583033 +0000 UTC m=+23.741239786 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.910048 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7rdsn" event={"ID":"bce814bc-030f-402e-ae0a-546f1c63edbd","Type":"ContainerStarted","Data":"53dab70a12bb25e3fa2611db860662c20975e597c052bc87ced71039bca9ff16"} Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.911808 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.913134 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2"} Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.913437 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.926597 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:27Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.941379 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:27Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.943947 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.943977 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.943998 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.944025 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:27 crc kubenswrapper[4745]: E1010 13:18:27.944086 4745 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 13:18:27 crc kubenswrapper[4745]: E1010 13:18:27.944148 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 13:18:27 crc kubenswrapper[4745]: E1010 13:18:27.944179 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 13:18:27 crc kubenswrapper[4745]: E1010 13:18:27.944177 4745 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 13:18:27 crc kubenswrapper[4745]: E1010 13:18:27.944215 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 13:18:27 crc kubenswrapper[4745]: E1010 13:18:27.944238 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 13:18:27 crc kubenswrapper[4745]: E1010 13:18:27.944187 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:29.944162727 +0000 UTC m=+23.841819560 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 13:18:27 crc kubenswrapper[4745]: E1010 13:18:27.944251 4745 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:27 crc kubenswrapper[4745]: E1010 13:18:27.944269 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:29.944252459 +0000 UTC m=+23.841909222 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 13:18:27 crc kubenswrapper[4745]: E1010 13:18:27.944191 4745 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:27 crc kubenswrapper[4745]: E1010 13:18:27.944293 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:29.944279839 +0000 UTC m=+23.841936602 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:27 crc kubenswrapper[4745]: E1010 13:18:27.944325 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:29.94431703 +0000 UTC m=+23.841973783 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.951641 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:27Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.965465 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:27Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:27 crc kubenswrapper[4745]: I1010 13:18:27.983604 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:27Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.006962 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.036838 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.072207 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.108893 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.262460 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-tzws7"] Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.262792 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-tzws7" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.264314 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.264529 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.266395 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.266880 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-4dnq2"] Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.267206 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-rlths"] Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.267369 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.268168 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-27d2n"] Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.268327 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5bh9m"] Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.268517 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.268870 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.269318 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.271107 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.271292 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.271402 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.271499 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.271621 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.271783 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.272203 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.272280 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.272397 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.272426 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.272538 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.272594 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.272780 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.272888 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.273284 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.274758 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.275623 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.275628 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.276872 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.288021 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.299413 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.314320 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.325854 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.336089 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.347719 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-os-release\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.347909 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-hostroot\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.348008 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-run-netns\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.348127 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a89f8c1b-e9aa-48a9-8072-254db014f8ef-cnibin\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.348237 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-var-lib-cni-bin\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.348334 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a89f8c1b-e9aa-48a9-8072-254db014f8ef-os-release\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.348430 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-cnibin\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.348538 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-run-netns\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.348626 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-multus-conf-dir\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.348748 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-node-log\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.348838 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-ovnkube-script-lib\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.348918 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-cni-binary-copy\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.348998 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/687bad4e-baff-4e09-bb33-def12c499092-ovn-node-metrics-cert\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.349069 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-run-k8s-cni-cncf-io\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.349143 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-var-lib-cni-multus\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.349210 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a89f8c1b-e9aa-48a9-8072-254db014f8ef-system-cni-dir\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.349273 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-env-overrides\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.349341 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c-rootfs\") pod \"machine-config-daemon-4dnq2\" (UID: \"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\") " pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.349430 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c-mcd-auth-proxy-config\") pod \"machine-config-daemon-4dnq2\" (UID: \"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\") " pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.349530 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6btr9\" (UniqueName: \"kubernetes.io/projected/a89f8c1b-e9aa-48a9-8072-254db014f8ef-kube-api-access-6btr9\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.349624 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-kubelet\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.349701 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-cni-bin\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.349815 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-var-lib-openvswitch\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.349896 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-ovn\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.349992 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-run-multus-certs\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.350289 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg98g\" (UniqueName: \"kubernetes.io/projected/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-kube-api-access-sg98g\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.350386 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-log-socket\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.350472 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9sbh\" (UniqueName: \"kubernetes.io/projected/687bad4e-baff-4e09-bb33-def12c499092-kube-api-access-l9sbh\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.350579 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-openvswitch\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.350674 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a89f8c1b-e9aa-48a9-8072-254db014f8ef-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.350818 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-systemd-units\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.350910 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-multus-cni-dir\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.351009 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-var-lib-kubelet\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.351108 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-multus-daemon-config\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.351201 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-slash\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.351294 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-cni-netd\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.351399 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-etc-kubernetes\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.351485 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a89f8c1b-e9aa-48a9-8072-254db014f8ef-cni-binary-copy\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.351583 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-system-cni-dir\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.351681 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-multus-socket-dir-parent\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.351801 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-etc-openvswitch\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.351899 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftf65\" (UniqueName: \"kubernetes.io/projected/5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c-kube-api-access-ftf65\") pod \"machine-config-daemon-4dnq2\" (UID: \"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\") " pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.352007 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn4tw\" (UniqueName: \"kubernetes.io/projected/9c112c93-ee29-4baa-aa3c-e370f4492bf0-kube-api-access-zn4tw\") pod \"node-resolver-tzws7\" (UID: \"9c112c93-ee29-4baa-aa3c-e370f4492bf0\") " pod="openshift-dns/node-resolver-tzws7" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.352131 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c-proxy-tls\") pod \"machine-config-daemon-4dnq2\" (UID: \"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\") " pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.352221 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a89f8c1b-e9aa-48a9-8072-254db014f8ef-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.352319 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.352421 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-ovnkube-config\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.352524 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9c112c93-ee29-4baa-aa3c-e370f4492bf0-hosts-file\") pod \"node-resolver-tzws7\" (UID: \"9c112c93-ee29-4baa-aa3c-e370f4492bf0\") " pod="openshift-dns/node-resolver-tzws7" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.352629 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-systemd\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.352759 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-run-ovn-kubernetes\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.357531 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.369564 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.381507 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.392031 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.403443 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.429742 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454360 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a89f8c1b-e9aa-48a9-8072-254db014f8ef-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454417 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454441 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-ovnkube-config\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454462 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c-proxy-tls\") pod \"machine-config-daemon-4dnq2\" (UID: \"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\") " pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454485 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9c112c93-ee29-4baa-aa3c-e370f4492bf0-hosts-file\") pod \"node-resolver-tzws7\" (UID: \"9c112c93-ee29-4baa-aa3c-e370f4492bf0\") " pod="openshift-dns/node-resolver-tzws7" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454503 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-systemd\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454522 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-run-ovn-kubernetes\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454555 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-hostroot\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454572 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-run-netns\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454591 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a89f8c1b-e9aa-48a9-8072-254db014f8ef-cnibin\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454611 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-os-release\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454630 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a89f8c1b-e9aa-48a9-8072-254db014f8ef-os-release\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454649 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-cnibin\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454670 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-var-lib-cni-bin\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454691 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-multus-conf-dir\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454711 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-node-log\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454751 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-ovnkube-script-lib\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454775 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-cni-binary-copy\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454797 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-run-netns\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454819 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/687bad4e-baff-4e09-bb33-def12c499092-ovn-node-metrics-cert\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454848 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-run-k8s-cni-cncf-io\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454869 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-var-lib-cni-multus\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454901 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a89f8c1b-e9aa-48a9-8072-254db014f8ef-system-cni-dir\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454922 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-env-overrides\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454950 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c-rootfs\") pod \"machine-config-daemon-4dnq2\" (UID: \"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\") " pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454969 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c-mcd-auth-proxy-config\") pod \"machine-config-daemon-4dnq2\" (UID: \"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\") " pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.454990 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6btr9\" (UniqueName: \"kubernetes.io/projected/a89f8c1b-e9aa-48a9-8072-254db014f8ef-kube-api-access-6btr9\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455010 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-kubelet\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455031 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-var-lib-openvswitch\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455049 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-ovn\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455070 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-cni-bin\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455094 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-run-multus-certs\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455116 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg98g\" (UniqueName: \"kubernetes.io/projected/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-kube-api-access-sg98g\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455134 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-log-socket\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455155 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9sbh\" (UniqueName: \"kubernetes.io/projected/687bad4e-baff-4e09-bb33-def12c499092-kube-api-access-l9sbh\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455178 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a89f8c1b-e9aa-48a9-8072-254db014f8ef-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455198 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-systemd-units\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455216 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-openvswitch\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455235 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-var-lib-kubelet\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455255 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-multus-daemon-config\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455276 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-slash\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455296 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-cni-netd\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455324 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455327 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-multus-cni-dir\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455402 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a89f8c1b-e9aa-48a9-8072-254db014f8ef-cni-binary-copy\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455426 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-system-cni-dir\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455448 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-etc-kubernetes\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455479 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-multus-socket-dir-parent\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455505 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-etc-openvswitch\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455524 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-log-socket\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455530 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftf65\" (UniqueName: \"kubernetes.io/projected/5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c-kube-api-access-ftf65\") pod \"machine-config-daemon-4dnq2\" (UID: \"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\") " pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455558 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn4tw\" (UniqueName: \"kubernetes.io/projected/9c112c93-ee29-4baa-aa3c-e370f4492bf0-kube-api-access-zn4tw\") pod \"node-resolver-tzws7\" (UID: \"9c112c93-ee29-4baa-aa3c-e370f4492bf0\") " pod="openshift-dns/node-resolver-tzws7" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455762 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-var-lib-cni-multus\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455794 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a89f8c1b-e9aa-48a9-8072-254db014f8ef-system-cni-dir\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455882 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-systemd-units\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455901 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-multus-cni-dir\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455937 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-openvswitch\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455940 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-var-lib-kubelet\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456040 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-kubelet\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456074 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c-rootfs\") pod \"machine-config-daemon-4dnq2\" (UID: \"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\") " pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456247 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-env-overrides\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456287 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-var-lib-openvswitch\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456309 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-ovn\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456332 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-cni-bin\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456354 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-run-multus-certs\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456355 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-ovnkube-config\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456490 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-node-log\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456508 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-multus-daemon-config\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456522 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9c112c93-ee29-4baa-aa3c-e370f4492bf0-hosts-file\") pod \"node-resolver-tzws7\" (UID: \"9c112c93-ee29-4baa-aa3c-e370f4492bf0\") " pod="openshift-dns/node-resolver-tzws7" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455251 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a89f8c1b-e9aa-48a9-8072-254db014f8ef-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456546 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-slash\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456555 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-systemd\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456569 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-cni-netd\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456575 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-run-ovn-kubernetes\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.455295 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-run-k8s-cni-cncf-io\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456724 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c-mcd-auth-proxy-config\") pod \"machine-config-daemon-4dnq2\" (UID: \"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\") " pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456806 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-os-release\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456813 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a89f8c1b-e9aa-48a9-8072-254db014f8ef-os-release\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456839 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-run-netns\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456837 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-hostroot\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456861 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a89f8c1b-e9aa-48a9-8072-254db014f8ef-cnibin\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456864 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-cnibin\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456902 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-etc-openvswitch\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456883 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-multus-conf-dir\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456879 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-var-lib-cni-bin\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456933 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-system-cni-dir\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456934 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-etc-kubernetes\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456969 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-multus-socket-dir-parent\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.456977 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a89f8c1b-e9aa-48a9-8072-254db014f8ef-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.457139 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-host-run-netns\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.457384 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-ovnkube-script-lib\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.457420 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-cni-binary-copy\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.457624 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a89f8c1b-e9aa-48a9-8072-254db014f8ef-cni-binary-copy\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.459458 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c-proxy-tls\") pod \"machine-config-daemon-4dnq2\" (UID: \"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\") " pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.466262 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/687bad4e-baff-4e09-bb33-def12c499092-ovn-node-metrics-cert\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.472004 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.482238 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn4tw\" (UniqueName: \"kubernetes.io/projected/9c112c93-ee29-4baa-aa3c-e370f4492bf0-kube-api-access-zn4tw\") pod \"node-resolver-tzws7\" (UID: \"9c112c93-ee29-4baa-aa3c-e370f4492bf0\") " pod="openshift-dns/node-resolver-tzws7" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.482339 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg98g\" (UniqueName: \"kubernetes.io/projected/1d62a30b-50a8-46a2-82c6-5d7e2f24e4de-kube-api-access-sg98g\") pod \"multus-27d2n\" (UID: \"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\") " pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.484617 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9sbh\" (UniqueName: \"kubernetes.io/projected/687bad4e-baff-4e09-bb33-def12c499092-kube-api-access-l9sbh\") pod \"ovnkube-node-5bh9m\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.489008 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftf65\" (UniqueName: \"kubernetes.io/projected/5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c-kube-api-access-ftf65\") pod \"machine-config-daemon-4dnq2\" (UID: \"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\") " pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.490125 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6btr9\" (UniqueName: \"kubernetes.io/projected/a89f8c1b-e9aa-48a9-8072-254db014f8ef-kube-api-access-6btr9\") pod \"multus-additional-cni-plugins-rlths\" (UID: \"a89f8c1b-e9aa-48a9-8072-254db014f8ef\") " pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.490402 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.504915 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.529196 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.551659 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.565511 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.575149 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.575174 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-tzws7" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.582620 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.587997 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.590705 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-27d2n" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.597994 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rlths" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.604498 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:28 crc kubenswrapper[4745]: W1010 13:18:28.607520 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d62a30b_50a8_46a2_82c6_5d7e2f24e4de.slice/crio-724ac8e63546da21d164c72c4b8b846b6c01124d99b924f56eda3f4eb3996b35 WatchSource:0}: Error finding container 724ac8e63546da21d164c72c4b8b846b6c01124d99b924f56eda3f4eb3996b35: Status 404 returned error can't find the container with id 724ac8e63546da21d164c72c4b8b846b6c01124d99b924f56eda3f4eb3996b35 Oct 10 13:18:28 crc kubenswrapper[4745]: W1010 13:18:28.610283 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5aaf7b57_cecd_4c6c_a45b_74504fa2dc2c.slice/crio-7f4bd1d04207f93fcabf0a56751bde831bf060a416117f851599a32145d14d5c WatchSource:0}: Error finding container 7f4bd1d04207f93fcabf0a56751bde831bf060a416117f851599a32145d14d5c: Status 404 returned error can't find the container with id 7f4bd1d04207f93fcabf0a56751bde831bf060a416117f851599a32145d14d5c Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.614219 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.633357 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.644191 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.661113 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.675151 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.744334 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:28 crc kubenswrapper[4745]: E1010 13:18:28.744701 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.744833 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:28 crc kubenswrapper[4745]: E1010 13:18:28.744929 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.917572 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-27d2n" event={"ID":"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de","Type":"ContainerStarted","Data":"9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb"} Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.917631 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-27d2n" event={"ID":"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de","Type":"ContainerStarted","Data":"724ac8e63546da21d164c72c4b8b846b6c01124d99b924f56eda3f4eb3996b35"} Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.919203 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1"} Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.920860 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5"} Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.920909 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb"} Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.920921 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"7f4bd1d04207f93fcabf0a56751bde831bf060a416117f851599a32145d14d5c"} Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.922526 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7rdsn" event={"ID":"bce814bc-030f-402e-ae0a-546f1c63edbd","Type":"ContainerStarted","Data":"eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981"} Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.923791 4745 generic.go:334] "Generic (PLEG): container finished" podID="687bad4e-baff-4e09-bb33-def12c499092" containerID="c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9" exitCode=0 Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.923849 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerDied","Data":"c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9"} Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.923872 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerStarted","Data":"23496822e4dfaba11f61f672b9fa2d403b2b98b040a78aaec8f472578b98952b"} Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.925684 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-tzws7" event={"ID":"9c112c93-ee29-4baa-aa3c-e370f4492bf0","Type":"ContainerStarted","Data":"b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb"} Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.925753 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-tzws7" event={"ID":"9c112c93-ee29-4baa-aa3c-e370f4492bf0","Type":"ContainerStarted","Data":"701137361dab2781b8de49d6e8dd82fc2e5ce0cc29e109e853849c571edb9343"} Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.927267 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" event={"ID":"a89f8c1b-e9aa-48a9-8072-254db014f8ef","Type":"ContainerStarted","Data":"7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e"} Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.927306 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" event={"ID":"a89f8c1b-e9aa-48a9-8072-254db014f8ef","Type":"ContainerStarted","Data":"1d0878db34792fc9df9428c287751b5574c9796ff1d70edfc70d24f966c0b023"} Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.933894 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.954689 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:28 crc kubenswrapper[4745]: I1010 13:18:28.989052 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:28Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.011007 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.062083 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.089964 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.108199 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.130021 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.142386 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.153080 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.161314 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.172073 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.188835 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.201831 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.210882 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.226528 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.235443 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.253280 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.272576 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.283433 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.293155 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.304362 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.319042 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.329991 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.342092 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.357777 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.369592 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.382286 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.743928 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:29 crc kubenswrapper[4745]: E1010 13:18:29.744060 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.753775 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.759389 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.761063 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.768762 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.785869 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.798291 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.817030 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.843560 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.855878 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.868224 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:18:29 crc kubenswrapper[4745]: E1010 13:18:29.868432 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:18:33.868396488 +0000 UTC m=+27.766053251 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.870459 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.884816 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.896752 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.906975 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.916821 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.929393 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.935440 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerStarted","Data":"79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210"} Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.935486 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerStarted","Data":"6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6"} Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.935499 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerStarted","Data":"5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1"} Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.935511 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerStarted","Data":"f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69"} Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.935523 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerStarted","Data":"982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9"} Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.935535 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerStarted","Data":"70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f"} Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.937847 4745 generic.go:334] "Generic (PLEG): container finished" podID="a89f8c1b-e9aa-48a9-8072-254db014f8ef" containerID="7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e" exitCode=0 Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.937982 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" event={"ID":"a89f8c1b-e9aa-48a9-8072-254db014f8ef","Type":"ContainerDied","Data":"7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e"} Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.945136 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.958843 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.969134 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.969181 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.969212 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.969240 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:29 crc kubenswrapper[4745]: E1010 13:18:29.969358 4745 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 13:18:29 crc kubenswrapper[4745]: E1010 13:18:29.969411 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:33.969395572 +0000 UTC m=+27.867052345 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 13:18:29 crc kubenswrapper[4745]: E1010 13:18:29.969480 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 13:18:29 crc kubenswrapper[4745]: E1010 13:18:29.969493 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 13:18:29 crc kubenswrapper[4745]: E1010 13:18:29.969507 4745 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:29 crc kubenswrapper[4745]: E1010 13:18:29.969534 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:33.969525425 +0000 UTC m=+27.867182198 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:29 crc kubenswrapper[4745]: E1010 13:18:29.969582 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 13:18:29 crc kubenswrapper[4745]: E1010 13:18:29.969593 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 13:18:29 crc kubenswrapper[4745]: E1010 13:18:29.969602 4745 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:29 crc kubenswrapper[4745]: E1010 13:18:29.969627 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:33.969619917 +0000 UTC m=+27.867276690 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:29 crc kubenswrapper[4745]: E1010 13:18:29.969660 4745 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 13:18:29 crc kubenswrapper[4745]: E1010 13:18:29.969682 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:33.969675349 +0000 UTC m=+27.867332122 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.971625 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.986845 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:29 crc kubenswrapper[4745]: I1010 13:18:29.999691 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:29Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.020822 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:30Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.035580 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:30Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.052088 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:30Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.073503 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:30Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.089473 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:30Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.103910 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:30Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.111949 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:30Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.122252 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:30Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.152510 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:30Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.193611 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:30Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.234958 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:30Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.276267 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:30Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.744599 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:30 crc kubenswrapper[4745]: E1010 13:18:30.745012 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.744613 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:30 crc kubenswrapper[4745]: E1010 13:18:30.745103 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.943021 4745 generic.go:334] "Generic (PLEG): container finished" podID="a89f8c1b-e9aa-48a9-8072-254db014f8ef" containerID="57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9" exitCode=0 Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.943067 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" event={"ID":"a89f8c1b-e9aa-48a9-8072-254db014f8ef","Type":"ContainerDied","Data":"57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9"} Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.953651 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:30Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.969492 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:30Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:30 crc kubenswrapper[4745]: I1010 13:18:30.985835 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:30Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.007412 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:31Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.022431 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:31Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.033063 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:31Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.044792 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:31Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.055295 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:31Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.074695 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:31Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.087542 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:31Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.098857 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:31Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.107553 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:31Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.120168 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:31Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.134690 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:31Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.149093 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:31Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.744534 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:31 crc kubenswrapper[4745]: E1010 13:18:31.744652 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.948257 4745 generic.go:334] "Generic (PLEG): container finished" podID="a89f8c1b-e9aa-48a9-8072-254db014f8ef" containerID="22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447" exitCode=0 Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.948327 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" event={"ID":"a89f8c1b-e9aa-48a9-8072-254db014f8ef","Type":"ContainerDied","Data":"22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447"} Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.951783 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerStarted","Data":"ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce"} Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.968165 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:31Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.987030 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:31Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:31 crc kubenswrapper[4745]: I1010 13:18:31.998828 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:31Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.011372 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.022844 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.037503 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.048141 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.063559 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.082364 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.096162 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.113487 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.125419 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.138077 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.151312 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.164282 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.615266 4745 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.618718 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.618786 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.618849 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.618982 4745 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.628383 4745 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.628803 4745 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.630155 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.630197 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.630214 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.630234 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.630251 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:32Z","lastTransitionTime":"2025-10-10T13:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:32 crc kubenswrapper[4745]: E1010 13:18:32.647156 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.653206 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.653234 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.653244 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.653260 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.653271 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:32Z","lastTransitionTime":"2025-10-10T13:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:32 crc kubenswrapper[4745]: E1010 13:18:32.676497 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.680889 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.680915 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.680923 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.680937 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.680950 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:32Z","lastTransitionTime":"2025-10-10T13:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:32 crc kubenswrapper[4745]: E1010 13:18:32.695127 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.698796 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.698887 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.698914 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.698947 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.699017 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:32Z","lastTransitionTime":"2025-10-10T13:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:32 crc kubenswrapper[4745]: E1010 13:18:32.719010 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.722309 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.722352 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.722365 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.722395 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.722409 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:32Z","lastTransitionTime":"2025-10-10T13:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:32 crc kubenswrapper[4745]: E1010 13:18:32.734072 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:32 crc kubenswrapper[4745]: E1010 13:18:32.734317 4745 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.735941 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.735970 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.735980 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.735995 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.736006 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:32Z","lastTransitionTime":"2025-10-10T13:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.745031 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:32 crc kubenswrapper[4745]: E1010 13:18:32.745292 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.745565 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:32 crc kubenswrapper[4745]: E1010 13:18:32.745695 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.839470 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.839512 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.839520 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.839535 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.839545 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:32Z","lastTransitionTime":"2025-10-10T13:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.941979 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.942021 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.942030 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.942047 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.942056 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:32Z","lastTransitionTime":"2025-10-10T13:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.958356 4745 generic.go:334] "Generic (PLEG): container finished" podID="a89f8c1b-e9aa-48a9-8072-254db014f8ef" containerID="0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95" exitCode=0 Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.958409 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" event={"ID":"a89f8c1b-e9aa-48a9-8072-254db014f8ef","Type":"ContainerDied","Data":"0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95"} Oct 10 13:18:32 crc kubenswrapper[4745]: I1010 13:18:32.985603 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:32Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.002892 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:33Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.020619 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:33Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.038826 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:33Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.045485 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.045538 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.045555 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.045579 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.045597 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:33Z","lastTransitionTime":"2025-10-10T13:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.055087 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:33Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.068340 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:33Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.078970 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:33Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.097709 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:33Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.111550 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:33Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.131357 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:33Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.145562 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:33Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.148297 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.148345 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.148357 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.148375 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.148389 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:33Z","lastTransitionTime":"2025-10-10T13:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.158703 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:33Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.173164 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:33Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.185530 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:33Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.196349 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:33Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.251392 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.251441 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.251454 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.251476 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.251490 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:33Z","lastTransitionTime":"2025-10-10T13:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.353980 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.354061 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.354090 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.354116 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.354141 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:33Z","lastTransitionTime":"2025-10-10T13:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.456642 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.456691 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.456703 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.456721 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.456753 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:33Z","lastTransitionTime":"2025-10-10T13:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.559177 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.559217 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.559229 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.559247 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.559259 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:33Z","lastTransitionTime":"2025-10-10T13:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.665815 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.665855 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.665864 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.665878 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.665888 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:33Z","lastTransitionTime":"2025-10-10T13:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.744648 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:33 crc kubenswrapper[4745]: E1010 13:18:33.744941 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.768916 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.768971 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.768991 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.769017 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.769036 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:33Z","lastTransitionTime":"2025-10-10T13:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.871768 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.871814 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.871829 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.871852 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.871870 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:33Z","lastTransitionTime":"2025-10-10T13:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.904886 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:18:33 crc kubenswrapper[4745]: E1010 13:18:33.905103 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:18:41.905076034 +0000 UTC m=+35.802732817 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.964545 4745 generic.go:334] "Generic (PLEG): container finished" podID="a89f8c1b-e9aa-48a9-8072-254db014f8ef" containerID="fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada" exitCode=0 Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.964586 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" event={"ID":"a89f8c1b-e9aa-48a9-8072-254db014f8ef","Type":"ContainerDied","Data":"fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada"} Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.973662 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.974242 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.974256 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.974273 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.974286 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:33Z","lastTransitionTime":"2025-10-10T13:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.979236 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:33Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:33 crc kubenswrapper[4745]: I1010 13:18:33.996331 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:33Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.006607 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.006661 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.006690 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.006719 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:34 crc kubenswrapper[4745]: E1010 13:18:34.006858 4745 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 13:18:34 crc kubenswrapper[4745]: E1010 13:18:34.006875 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 13:18:34 crc kubenswrapper[4745]: E1010 13:18:34.006905 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 13:18:34 crc kubenswrapper[4745]: E1010 13:18:34.006920 4745 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:34 crc kubenswrapper[4745]: E1010 13:18:34.006921 4745 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 13:18:34 crc kubenswrapper[4745]: E1010 13:18:34.006940 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 13:18:34 crc kubenswrapper[4745]: E1010 13:18:34.006981 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 13:18:34 crc kubenswrapper[4745]: E1010 13:18:34.006998 4745 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:34 crc kubenswrapper[4745]: E1010 13:18:34.006950 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:42.00692165 +0000 UTC m=+35.904578423 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 13:18:34 crc kubenswrapper[4745]: E1010 13:18:34.007122 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:42.007080894 +0000 UTC m=+35.904737807 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:34 crc kubenswrapper[4745]: E1010 13:18:34.007143 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:42.007134145 +0000 UTC m=+35.904790908 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 13:18:34 crc kubenswrapper[4745]: E1010 13:18:34.007161 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:42.007151826 +0000 UTC m=+35.904808819 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.011123 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.023609 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.038674 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.064227 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.078510 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.078603 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.078622 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.078656 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.078678 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:34Z","lastTransitionTime":"2025-10-10T13:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.087248 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.102032 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.114701 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.127919 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.162046 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.184102 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.184151 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.184160 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.184174 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.184184 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:34Z","lastTransitionTime":"2025-10-10T13:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.184694 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.206569 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.221491 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.232671 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.285967 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.286024 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.286035 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.286059 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.286115 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:34Z","lastTransitionTime":"2025-10-10T13:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.390581 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.390648 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.390673 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.390702 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.390793 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:34Z","lastTransitionTime":"2025-10-10T13:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.493898 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.493957 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.493967 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.493986 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.493999 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:34Z","lastTransitionTime":"2025-10-10T13:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.596182 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.596222 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.596233 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.596247 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.596259 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:34Z","lastTransitionTime":"2025-10-10T13:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.699618 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.699684 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.699694 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.699709 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.699717 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:34Z","lastTransitionTime":"2025-10-10T13:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.744254 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:34 crc kubenswrapper[4745]: E1010 13:18:34.744377 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.744387 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:34 crc kubenswrapper[4745]: E1010 13:18:34.744602 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.802816 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.802895 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.802914 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.802952 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.802986 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:34Z","lastTransitionTime":"2025-10-10T13:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.905660 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.905707 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.905718 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.905752 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.905764 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:34Z","lastTransitionTime":"2025-10-10T13:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.974538 4745 generic.go:334] "Generic (PLEG): container finished" podID="a89f8c1b-e9aa-48a9-8072-254db014f8ef" containerID="3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029" exitCode=0 Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.974630 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" event={"ID":"a89f8c1b-e9aa-48a9-8072-254db014f8ef","Type":"ContainerDied","Data":"3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029"} Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.982913 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerStarted","Data":"6c9d489fb09997c3527a01e00688369ff7301ca7220da398ee334873d64a6449"} Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.983356 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:34 crc kubenswrapper[4745]: I1010 13:18:34.996705 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.009036 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.009077 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.009088 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.009103 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.009116 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:35Z","lastTransitionTime":"2025-10-10T13:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.011844 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.024058 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.027765 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.040831 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.055099 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.069388 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.088516 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.100606 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.111242 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.111240 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.111271 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.111279 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.111291 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.111301 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:35Z","lastTransitionTime":"2025-10-10T13:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.123139 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.133202 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.152519 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.163612 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.175604 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.186882 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.198645 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.209697 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.213623 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.213719 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.213862 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.213940 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.213967 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:35Z","lastTransitionTime":"2025-10-10T13:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.221169 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.239395 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9d489fb09997c3527a01e00688369ff7301ca7220da398ee334873d64a6449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.256362 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.273641 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.285097 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.296891 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.309996 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.317135 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.317244 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.317315 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.317390 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.317458 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:35Z","lastTransitionTime":"2025-10-10T13:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.323088 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.334387 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.346068 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.367204 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.379404 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.395282 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:35Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.419917 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.420016 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.420085 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.420151 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.420215 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:35Z","lastTransitionTime":"2025-10-10T13:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.523614 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.524027 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.524247 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.524404 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.524537 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:35Z","lastTransitionTime":"2025-10-10T13:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.627755 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.627810 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.627829 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.627851 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.627866 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:35Z","lastTransitionTime":"2025-10-10T13:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.730977 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.731041 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.731060 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.731084 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.731101 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:35Z","lastTransitionTime":"2025-10-10T13:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.744252 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:35 crc kubenswrapper[4745]: E1010 13:18:35.744404 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.833932 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.834347 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.834502 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.834680 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.834919 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:35Z","lastTransitionTime":"2025-10-10T13:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.937958 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.938022 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.938040 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.938063 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.938081 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:35Z","lastTransitionTime":"2025-10-10T13:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.990991 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" event={"ID":"a89f8c1b-e9aa-48a9-8072-254db014f8ef","Type":"ContainerStarted","Data":"35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b"} Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.991119 4745 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 10 13:18:35 crc kubenswrapper[4745]: I1010 13:18:35.991838 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.009563 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.026641 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.031074 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.040845 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.040916 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.040932 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.040962 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.041006 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:36Z","lastTransitionTime":"2025-10-10T13:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.051587 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.075673 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.096442 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.122166 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.142240 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.143584 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.143672 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.143695 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.143725 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.143776 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:36Z","lastTransitionTime":"2025-10-10T13:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.162591 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.180400 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.200456 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.217865 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.242631 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9d489fb09997c3527a01e00688369ff7301ca7220da398ee334873d64a6449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.247084 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.247177 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.247197 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.247222 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.247240 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:36Z","lastTransitionTime":"2025-10-10T13:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.262986 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.295352 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.315604 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.334580 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.349283 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.349608 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.349644 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.349662 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.349687 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.349707 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:36Z","lastTransitionTime":"2025-10-10T13:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.366545 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.387694 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.407646 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.425408 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.438554 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.456505 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.456560 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.456572 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.456610 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.456622 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:36Z","lastTransitionTime":"2025-10-10T13:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.458841 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.478822 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.505717 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.525354 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.540495 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.557913 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.559401 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.559440 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.559454 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.559471 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.559483 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:36Z","lastTransitionTime":"2025-10-10T13:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.572983 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.595321 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9d489fb09997c3527a01e00688369ff7301ca7220da398ee334873d64a6449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.662182 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.662245 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.662261 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.662284 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.662302 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:36Z","lastTransitionTime":"2025-10-10T13:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.744311 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:36 crc kubenswrapper[4745]: E1010 13:18:36.744441 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.744557 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:36 crc kubenswrapper[4745]: E1010 13:18:36.744805 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.757316 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.764822 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.764853 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.764863 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.764876 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.764886 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:36Z","lastTransitionTime":"2025-10-10T13:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.768213 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.779799 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.791285 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.801872 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.812664 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.828780 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.840796 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.859126 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.867245 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.867297 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.867306 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.867341 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.867352 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:36Z","lastTransitionTime":"2025-10-10T13:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.871109 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.882919 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.895109 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.916479 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9d489fb09997c3527a01e00688369ff7301ca7220da398ee334873d64a6449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.930545 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.952013 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.969680 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.969721 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.969742 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.969790 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.969801 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:36Z","lastTransitionTime":"2025-10-10T13:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:36 crc kubenswrapper[4745]: I1010 13:18:36.994823 4745 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.072528 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.072566 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.072574 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.072589 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.072600 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:37Z","lastTransitionTime":"2025-10-10T13:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.175297 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.175366 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.175392 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.175423 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.175447 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:37Z","lastTransitionTime":"2025-10-10T13:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.277474 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.277506 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.277514 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.277528 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.277538 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:37Z","lastTransitionTime":"2025-10-10T13:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.379970 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.380013 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.380028 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.380048 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.380062 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:37Z","lastTransitionTime":"2025-10-10T13:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.483201 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.483250 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.483265 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.483285 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.483299 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:37Z","lastTransitionTime":"2025-10-10T13:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.586093 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.586126 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.586137 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.586153 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.586165 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:37Z","lastTransitionTime":"2025-10-10T13:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.689140 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.689175 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.689185 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.689244 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.689261 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:37Z","lastTransitionTime":"2025-10-10T13:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.744130 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:37 crc kubenswrapper[4745]: E1010 13:18:37.744255 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.792748 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.792819 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.792833 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.792849 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.793197 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:37Z","lastTransitionTime":"2025-10-10T13:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.896521 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.896608 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.896640 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.896675 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.896703 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:37Z","lastTransitionTime":"2025-10-10T13:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.998476 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.998509 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.998517 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.998528 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:37 crc kubenswrapper[4745]: I1010 13:18:37.998538 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:37Z","lastTransitionTime":"2025-10-10T13:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.000368 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovnkube-controller/0.log" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.003971 4745 generic.go:334] "Generic (PLEG): container finished" podID="687bad4e-baff-4e09-bb33-def12c499092" containerID="6c9d489fb09997c3527a01e00688369ff7301ca7220da398ee334873d64a6449" exitCode=1 Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.004029 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerDied","Data":"6c9d489fb09997c3527a01e00688369ff7301ca7220da398ee334873d64a6449"} Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.004462 4745 scope.go:117] "RemoveContainer" containerID="6c9d489fb09997c3527a01e00688369ff7301ca7220da398ee334873d64a6449" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.028362 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.045691 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.065280 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.079245 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.100567 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.100601 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.100611 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.100626 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.100637 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:38Z","lastTransitionTime":"2025-10-10T13:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.106929 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9d489fb09997c3527a01e00688369ff7301ca7220da398ee334873d64a6449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c9d489fb09997c3527a01e00688369ff7301ca7220da398ee334873d64a6449\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:37Z\\\",\\\"message\\\":\\\"d (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 13:18:37.264282 5998 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1010 13:18:37.264304 5998 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1010 13:18:37.264321 5998 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 13:18:37.264327 5998 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1010 13:18:37.264346 5998 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1010 13:18:37.264365 5998 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 13:18:37.264381 5998 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1010 13:18:37.264390 5998 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1010 13:18:37.264407 5998 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1010 13:18:37.264431 5998 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1010 13:18:37.264439 5998 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 13:18:37.264449 5998 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1010 13:18:37.264457 5998 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 13:18:37.264475 5998 factory.go:656] Stopping watch factory\\\\nI1010 13:18:37.264495 5998 ovnkube.go:599] Stopped ovnkube\\\\nI1010 13:18:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.122915 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.145614 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.161540 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.172442 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.192026 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.203056 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.203094 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.203108 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.203124 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.203136 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:38Z","lastTransitionTime":"2025-10-10T13:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.210435 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.226320 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.240509 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.252306 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.266661 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.305993 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.306030 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.306040 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.306056 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.306066 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:38Z","lastTransitionTime":"2025-10-10T13:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.383391 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.395859 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.407623 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.407667 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.407677 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.407693 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.407704 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:38Z","lastTransitionTime":"2025-10-10T13:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.412743 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.433215 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9d489fb09997c3527a01e00688369ff7301ca7220da398ee334873d64a6449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c9d489fb09997c3527a01e00688369ff7301ca7220da398ee334873d64a6449\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:37Z\\\",\\\"message\\\":\\\"d (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 13:18:37.264282 5998 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1010 13:18:37.264304 5998 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1010 13:18:37.264321 5998 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 13:18:37.264327 5998 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1010 13:18:37.264346 5998 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1010 13:18:37.264365 5998 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 13:18:37.264381 5998 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1010 13:18:37.264390 5998 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1010 13:18:37.264407 5998 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1010 13:18:37.264431 5998 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1010 13:18:37.264439 5998 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 13:18:37.264449 5998 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1010 13:18:37.264457 5998 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 13:18:37.264475 5998 factory.go:656] Stopping watch factory\\\\nI1010 13:18:37.264495 5998 ovnkube.go:599] Stopped ovnkube\\\\nI1010 13:18:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.448630 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.472134 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.485950 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.504594 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.510306 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.510354 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.510367 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.510383 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.510572 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:38Z","lastTransitionTime":"2025-10-10T13:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.518178 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.528535 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.540278 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.551098 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.559975 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.570872 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.584122 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.596511 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.612799 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.612840 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.612848 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.612863 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.612880 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:38Z","lastTransitionTime":"2025-10-10T13:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.714929 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.714968 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.714976 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.714989 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.714998 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:38Z","lastTransitionTime":"2025-10-10T13:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.744612 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.744683 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:38 crc kubenswrapper[4745]: E1010 13:18:38.744777 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:38 crc kubenswrapper[4745]: E1010 13:18:38.744887 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.817467 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.817522 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.817535 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.817556 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.817571 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:38Z","lastTransitionTime":"2025-10-10T13:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.920358 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.920402 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.920411 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.920429 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:38 crc kubenswrapper[4745]: I1010 13:18:38.920445 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:38Z","lastTransitionTime":"2025-10-10T13:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.011961 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovnkube-controller/1.log" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.012909 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovnkube-controller/0.log" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.017664 4745 generic.go:334] "Generic (PLEG): container finished" podID="687bad4e-baff-4e09-bb33-def12c499092" containerID="c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52" exitCode=1 Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.017778 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerDied","Data":"c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52"} Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.017923 4745 scope.go:117] "RemoveContainer" containerID="6c9d489fb09997c3527a01e00688369ff7301ca7220da398ee334873d64a6449" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.018473 4745 scope.go:117] "RemoveContainer" containerID="c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52" Oct 10 13:18:39 crc kubenswrapper[4745]: E1010 13:18:39.018666 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podUID="687bad4e-baff-4e09-bb33-def12c499092" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.023638 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.023685 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.023697 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.023716 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.023750 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:39Z","lastTransitionTime":"2025-10-10T13:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.037461 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:39Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.062346 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:39Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.087678 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:39Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.110980 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:39Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.133058 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.133110 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.133120 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.133139 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.133149 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:39Z","lastTransitionTime":"2025-10-10T13:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.134051 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:39Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.145549 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:39Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.169907 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c9d489fb09997c3527a01e00688369ff7301ca7220da398ee334873d64a6449\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:37Z\\\",\\\"message\\\":\\\"d (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 13:18:37.264282 5998 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1010 13:18:37.264304 5998 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1010 13:18:37.264321 5998 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 13:18:37.264327 5998 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1010 13:18:37.264346 5998 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1010 13:18:37.264365 5998 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 13:18:37.264381 5998 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1010 13:18:37.264390 5998 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1010 13:18:37.264407 5998 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1010 13:18:37.264431 5998 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1010 13:18:37.264439 5998 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 13:18:37.264449 5998 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1010 13:18:37.264457 5998 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 13:18:37.264475 5998 factory.go:656] Stopping watch factory\\\\nI1010 13:18:37.264495 5998 ovnkube.go:599] Stopped ovnkube\\\\nI1010 13:18:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"message\\\":\\\"de network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z]\\\\nI1010 13:18:38.877288 6158 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-7rdsn after 0 failed attempt(s)\\\\nI1010 13:18:38.877300 6158 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7rdsn\\\\nI1010 13:18:38.876514 6158 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1010 13:18:38.876469 6158 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:39Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.184852 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:39Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.205458 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:39Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.219176 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:39Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.231336 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:39Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.236126 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.236195 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.236206 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.236224 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.236236 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:39Z","lastTransitionTime":"2025-10-10T13:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.245159 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:39Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.267796 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:39Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.283924 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:39Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.300185 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:39Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.338803 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.338850 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.338863 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.338887 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.338920 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:39Z","lastTransitionTime":"2025-10-10T13:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.442641 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.442765 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.442788 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.442827 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.442882 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:39Z","lastTransitionTime":"2025-10-10T13:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.546152 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.546224 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.546242 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.546268 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.546288 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:39Z","lastTransitionTime":"2025-10-10T13:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.649624 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.649695 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.649713 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.649767 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.649785 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:39Z","lastTransitionTime":"2025-10-10T13:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.744410 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:39 crc kubenswrapper[4745]: E1010 13:18:39.744610 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.754621 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.754687 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.754706 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.754782 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.754806 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:39Z","lastTransitionTime":"2025-10-10T13:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.858084 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.858147 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.858167 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.858195 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.858213 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:39Z","lastTransitionTime":"2025-10-10T13:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.962316 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.962383 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.962403 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.962431 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:39 crc kubenswrapper[4745]: I1010 13:18:39.962447 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:39Z","lastTransitionTime":"2025-10-10T13:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.024637 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovnkube-controller/1.log" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.030496 4745 scope.go:117] "RemoveContainer" containerID="c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52" Oct 10 13:18:40 crc kubenswrapper[4745]: E1010 13:18:40.030910 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podUID="687bad4e-baff-4e09-bb33-def12c499092" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.051581 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.071507 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.071560 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.071579 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.071603 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.071621 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:40Z","lastTransitionTime":"2025-10-10T13:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.076274 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.094216 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.113024 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.133405 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.149785 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.174924 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.175193 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.175281 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.175370 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.175460 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:40Z","lastTransitionTime":"2025-10-10T13:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.187922 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"message\\\":\\\"de network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z]\\\\nI1010 13:18:38.877288 6158 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-7rdsn after 0 failed attempt(s)\\\\nI1010 13:18:38.877300 6158 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7rdsn\\\\nI1010 13:18:38.876514 6158 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1010 13:18:38.876469 6158 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.209619 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.242946 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.263879 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.279475 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.279529 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.279546 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.279574 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.279591 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:40Z","lastTransitionTime":"2025-10-10T13:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.280424 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.302920 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.323923 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.342385 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.360404 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.388685 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.389109 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.389229 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.389356 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.389483 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:40Z","lastTransitionTime":"2025-10-10T13:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.491980 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.492048 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.492066 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.492088 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.492105 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:40Z","lastTransitionTime":"2025-10-10T13:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.553356 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv"] Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.554069 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.556970 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.557810 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.573781 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.595514 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.595570 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.595588 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.595616 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.595638 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:40Z","lastTransitionTime":"2025-10-10T13:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.596356 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.615104 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.643718 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"message\\\":\\\"de network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z]\\\\nI1010 13:18:38.877288 6158 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-7rdsn after 0 failed attempt(s)\\\\nI1010 13:18:38.877300 6158 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7rdsn\\\\nI1010 13:18:38.876514 6158 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1010 13:18:38.876469 6158 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.665115 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.675932 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b60d4139-055c-4d5a-9529-1c81f5639053-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-62rnv\" (UID: \"b60d4139-055c-4d5a-9529-1c81f5639053\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.675992 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b60d4139-055c-4d5a-9529-1c81f5639053-env-overrides\") pod \"ovnkube-control-plane-749d76644c-62rnv\" (UID: \"b60d4139-055c-4d5a-9529-1c81f5639053\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.676064 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b60d4139-055c-4d5a-9529-1c81f5639053-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-62rnv\" (UID: \"b60d4139-055c-4d5a-9529-1c81f5639053\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.676129 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stvrc\" (UniqueName: \"kubernetes.io/projected/b60d4139-055c-4d5a-9529-1c81f5639053-kube-api-access-stvrc\") pod \"ovnkube-control-plane-749d76644c-62rnv\" (UID: \"b60d4139-055c-4d5a-9529-1c81f5639053\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.695251 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.697894 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.697996 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.698018 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.698047 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.698068 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:40Z","lastTransitionTime":"2025-10-10T13:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.713452 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.729403 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.744449 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.744632 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:40 crc kubenswrapper[4745]: E1010 13:18:40.744793 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:40 crc kubenswrapper[4745]: E1010 13:18:40.745005 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.747625 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.764921 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.776916 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stvrc\" (UniqueName: \"kubernetes.io/projected/b60d4139-055c-4d5a-9529-1c81f5639053-kube-api-access-stvrc\") pod \"ovnkube-control-plane-749d76644c-62rnv\" (UID: \"b60d4139-055c-4d5a-9529-1c81f5639053\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.777020 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b60d4139-055c-4d5a-9529-1c81f5639053-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-62rnv\" (UID: \"b60d4139-055c-4d5a-9529-1c81f5639053\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.777075 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b60d4139-055c-4d5a-9529-1c81f5639053-env-overrides\") pod \"ovnkube-control-plane-749d76644c-62rnv\" (UID: \"b60d4139-055c-4d5a-9529-1c81f5639053\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.777156 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b60d4139-055c-4d5a-9529-1c81f5639053-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-62rnv\" (UID: \"b60d4139-055c-4d5a-9529-1c81f5639053\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.778012 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b60d4139-055c-4d5a-9529-1c81f5639053-env-overrides\") pod \"ovnkube-control-plane-749d76644c-62rnv\" (UID: \"b60d4139-055c-4d5a-9529-1c81f5639053\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.778724 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b60d4139-055c-4d5a-9529-1c81f5639053-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-62rnv\" (UID: \"b60d4139-055c-4d5a-9529-1c81f5639053\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.781556 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.786659 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b60d4139-055c-4d5a-9529-1c81f5639053-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-62rnv\" (UID: \"b60d4139-055c-4d5a-9529-1c81f5639053\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.801324 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.801370 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.801383 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.801404 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.801417 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:40Z","lastTransitionTime":"2025-10-10T13:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.804671 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.806060 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stvrc\" (UniqueName: \"kubernetes.io/projected/b60d4139-055c-4d5a-9529-1c81f5639053-kube-api-access-stvrc\") pod \"ovnkube-control-plane-749d76644c-62rnv\" (UID: \"b60d4139-055c-4d5a-9529-1c81f5639053\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.818691 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.840951 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.861256 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.870517 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.882483 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:40Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:40 crc kubenswrapper[4745]: W1010 13:18:40.893860 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb60d4139_055c_4d5a_9529_1c81f5639053.slice/crio-cb03ce1988b45661a7a07c7943ec9148eed217c96b89b38876ed0d6d098807a2 WatchSource:0}: Error finding container cb03ce1988b45661a7a07c7943ec9148eed217c96b89b38876ed0d6d098807a2: Status 404 returned error can't find the container with id cb03ce1988b45661a7a07c7943ec9148eed217c96b89b38876ed0d6d098807a2 Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.905416 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.905477 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.905497 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.905524 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:40 crc kubenswrapper[4745]: I1010 13:18:40.905543 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:40Z","lastTransitionTime":"2025-10-10T13:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.008940 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.008994 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.009006 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.009027 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.009040 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:41Z","lastTransitionTime":"2025-10-10T13:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.033812 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" event={"ID":"b60d4139-055c-4d5a-9529-1c81f5639053","Type":"ContainerStarted","Data":"cb03ce1988b45661a7a07c7943ec9148eed217c96b89b38876ed0d6d098807a2"} Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.112010 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.112054 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.112065 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.112086 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.112099 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:41Z","lastTransitionTime":"2025-10-10T13:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.215942 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.215998 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.216017 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.216046 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.216066 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:41Z","lastTransitionTime":"2025-10-10T13:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.319285 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.319356 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.319375 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.319405 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.319426 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:41Z","lastTransitionTime":"2025-10-10T13:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.423284 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.423358 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.423372 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.423400 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.423416 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:41Z","lastTransitionTime":"2025-10-10T13:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.529254 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.529529 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.529593 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.529660 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.529753 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:41Z","lastTransitionTime":"2025-10-10T13:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.633159 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.633250 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.633269 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.633298 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.633319 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:41Z","lastTransitionTime":"2025-10-10T13:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.737094 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.737146 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.737158 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.737180 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.737192 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:41Z","lastTransitionTime":"2025-10-10T13:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.744807 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:41 crc kubenswrapper[4745]: E1010 13:18:41.745034 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.840812 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.840858 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.840871 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.840891 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.840906 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:41Z","lastTransitionTime":"2025-10-10T13:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.943927 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.944409 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.944588 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.944759 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.944894 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:41Z","lastTransitionTime":"2025-10-10T13:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:41 crc kubenswrapper[4745]: I1010 13:18:41.992088 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:18:41 crc kubenswrapper[4745]: E1010 13:18:41.992385 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:18:57.992341496 +0000 UTC m=+51.889998299 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.040313 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" event={"ID":"b60d4139-055c-4d5a-9529-1c81f5639053","Type":"ContainerStarted","Data":"ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3"} Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.040408 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" event={"ID":"b60d4139-055c-4d5a-9529-1c81f5639053","Type":"ContainerStarted","Data":"d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113"} Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.047293 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.047336 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.047354 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.047380 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.047398 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:42Z","lastTransitionTime":"2025-10-10T13:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.064019 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.085856 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.094399 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.094650 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.094772 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.094828 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.094858 4745 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.094969 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.095022 4745 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.095053 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.095082 4745 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.094980 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.095012 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:58.094989851 +0000 UTC m=+51.992646614 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.095613 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:58.095582406 +0000 UTC m=+51.993239219 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.095658 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:58.095634017 +0000 UTC m=+51.993291020 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.095702 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.096005 4745 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.096099 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:58.096073788 +0000 UTC m=+51.993730591 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.107574 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.123312 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-xgh8p"] Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.124327 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.124454 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.127316 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.150566 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.150652 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.150680 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.150717 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.150785 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:42Z","lastTransitionTime":"2025-10-10T13:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.158253 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"message\\\":\\\"de network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z]\\\\nI1010 13:18:38.877288 6158 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-7rdsn after 0 failed attempt(s)\\\\nI1010 13:18:38.877300 6158 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7rdsn\\\\nI1010 13:18:38.876514 6158 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1010 13:18:38.876469 6158 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.180319 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.196944 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrj8j\" (UniqueName: \"kubernetes.io/projected/82c4e63a-009f-4720-87b5-640a8ec65117-kube-api-access-qrj8j\") pod \"network-metrics-daemon-xgh8p\" (UID: \"82c4e63a-009f-4720-87b5-640a8ec65117\") " pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.197025 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs\") pod \"network-metrics-daemon-xgh8p\" (UID: \"82c4e63a-009f-4720-87b5-640a8ec65117\") " pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.208475 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.230844 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.249120 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.253647 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.253710 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.253766 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.253800 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.253822 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:42Z","lastTransitionTime":"2025-10-10T13:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.271386 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.294655 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.299149 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrj8j\" (UniqueName: \"kubernetes.io/projected/82c4e63a-009f-4720-87b5-640a8ec65117-kube-api-access-qrj8j\") pod \"network-metrics-daemon-xgh8p\" (UID: \"82c4e63a-009f-4720-87b5-640a8ec65117\") " pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.299280 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs\") pod \"network-metrics-daemon-xgh8p\" (UID: \"82c4e63a-009f-4720-87b5-640a8ec65117\") " pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.299497 4745 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.299626 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs podName:82c4e63a-009f-4720-87b5-640a8ec65117 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:42.799591145 +0000 UTC m=+36.697247948 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs") pod "network-metrics-daemon-xgh8p" (UID: "82c4e63a-009f-4720-87b5-640a8ec65117") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.315681 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.326925 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrj8j\" (UniqueName: \"kubernetes.io/projected/82c4e63a-009f-4720-87b5-640a8ec65117-kube-api-access-qrj8j\") pod \"network-metrics-daemon-xgh8p\" (UID: \"82c4e63a-009f-4720-87b5-640a8ec65117\") " pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.334985 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.357095 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.357187 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.357170 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.357216 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.357489 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.357521 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:42Z","lastTransitionTime":"2025-10-10T13:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.375524 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.396698 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.418776 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.436063 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.454371 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82c4e63a-009f-4720-87b5-640a8ec65117\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xgh8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.460642 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.460709 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.460757 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.460785 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.460807 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:42Z","lastTransitionTime":"2025-10-10T13:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.474451 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.495143 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.514206 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.533476 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.550559 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.565835 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.565895 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.565909 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.565930 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.565944 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:42Z","lastTransitionTime":"2025-10-10T13:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.579743 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"message\\\":\\\"de network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z]\\\\nI1010 13:18:38.877288 6158 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-7rdsn after 0 failed attempt(s)\\\\nI1010 13:18:38.877300 6158 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7rdsn\\\\nI1010 13:18:38.876514 6158 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1010 13:18:38.876469 6158 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.599546 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.632245 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.654842 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.669439 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.669524 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.669547 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.669601 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.669627 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:42Z","lastTransitionTime":"2025-10-10T13:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.673419 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.695102 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.720381 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.743983 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.744080 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.744189 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.744274 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.745883 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.768115 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:42Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.772901 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.772958 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.772971 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.772994 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.773011 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:42Z","lastTransitionTime":"2025-10-10T13:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.806315 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs\") pod \"network-metrics-daemon-xgh8p\" (UID: \"82c4e63a-009f-4720-87b5-640a8ec65117\") " pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.806703 4745 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 13:18:42 crc kubenswrapper[4745]: E1010 13:18:42.807092 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs podName:82c4e63a-009f-4720-87b5-640a8ec65117 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:43.807057268 +0000 UTC m=+37.704714061 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs") pod "network-metrics-daemon-xgh8p" (UID: "82c4e63a-009f-4720-87b5-640a8ec65117") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.878291 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.878484 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.878506 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.878543 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.878568 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:42Z","lastTransitionTime":"2025-10-10T13:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.982693 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.982795 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.982819 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.982849 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:42 crc kubenswrapper[4745]: I1010 13:18:42.982872 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:42Z","lastTransitionTime":"2025-10-10T13:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.085920 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.086893 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.087249 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.087578 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.087651 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:43Z","lastTransitionTime":"2025-10-10T13:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.129710 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.129822 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.129842 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.130285 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.130345 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:43Z","lastTransitionTime":"2025-10-10T13:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:43 crc kubenswrapper[4745]: E1010 13:18:43.150602 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:43Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.157535 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.157604 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.157632 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.157667 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.157691 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:43Z","lastTransitionTime":"2025-10-10T13:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:43 crc kubenswrapper[4745]: E1010 13:18:43.181517 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:43Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.187624 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.187787 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.187818 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.187897 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.187925 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:43Z","lastTransitionTime":"2025-10-10T13:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:43 crc kubenswrapper[4745]: E1010 13:18:43.211427 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:43Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.218622 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.218673 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.218692 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.218755 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.218775 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:43Z","lastTransitionTime":"2025-10-10T13:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:43 crc kubenswrapper[4745]: E1010 13:18:43.238914 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:43Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.245279 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.245389 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.245444 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.245473 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.245720 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:43Z","lastTransitionTime":"2025-10-10T13:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:43 crc kubenswrapper[4745]: E1010 13:18:43.269083 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:43Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:43 crc kubenswrapper[4745]: E1010 13:18:43.269374 4745 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.271910 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.271947 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.271964 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.271990 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.272010 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:43Z","lastTransitionTime":"2025-10-10T13:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.375748 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.375811 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.375825 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.375848 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.375863 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:43Z","lastTransitionTime":"2025-10-10T13:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.480035 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.480094 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.480107 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.480133 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.480151 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:43Z","lastTransitionTime":"2025-10-10T13:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.584140 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.584218 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.584234 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.584260 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.584276 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:43Z","lastTransitionTime":"2025-10-10T13:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.687667 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.688207 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.688296 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.688437 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.688522 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:43Z","lastTransitionTime":"2025-10-10T13:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.744326 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.744358 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:43 crc kubenswrapper[4745]: E1010 13:18:43.744795 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:18:43 crc kubenswrapper[4745]: E1010 13:18:43.744801 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.790807 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.790868 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.790884 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.790905 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.790920 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:43Z","lastTransitionTime":"2025-10-10T13:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.821341 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs\") pod \"network-metrics-daemon-xgh8p\" (UID: \"82c4e63a-009f-4720-87b5-640a8ec65117\") " pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:43 crc kubenswrapper[4745]: E1010 13:18:43.821479 4745 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 13:18:43 crc kubenswrapper[4745]: E1010 13:18:43.821683 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs podName:82c4e63a-009f-4720-87b5-640a8ec65117 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:45.821667845 +0000 UTC m=+39.719324608 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs") pod "network-metrics-daemon-xgh8p" (UID: "82c4e63a-009f-4720-87b5-640a8ec65117") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.892884 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.893118 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.893198 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.893286 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.893366 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:43Z","lastTransitionTime":"2025-10-10T13:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.995354 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.995404 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.995420 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.995442 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:43 crc kubenswrapper[4745]: I1010 13:18:43.995459 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:43Z","lastTransitionTime":"2025-10-10T13:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.098747 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.098788 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.098797 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.098813 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.098825 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:44Z","lastTransitionTime":"2025-10-10T13:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.202300 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.202378 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.202396 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.202422 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.202440 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:44Z","lastTransitionTime":"2025-10-10T13:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.305534 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.305589 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.305604 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.305622 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.305636 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:44Z","lastTransitionTime":"2025-10-10T13:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.409103 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.409158 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.409175 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.409199 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.409219 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:44Z","lastTransitionTime":"2025-10-10T13:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.512825 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.512867 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.512880 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.512900 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.512914 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:44Z","lastTransitionTime":"2025-10-10T13:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.615605 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.615654 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.615669 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.615686 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.615698 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:44Z","lastTransitionTime":"2025-10-10T13:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.719181 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.719250 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.719277 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.719313 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.719339 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:44Z","lastTransitionTime":"2025-10-10T13:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.745048 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.745384 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:44 crc kubenswrapper[4745]: E1010 13:18:44.745635 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:44 crc kubenswrapper[4745]: E1010 13:18:44.745665 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.822292 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.822370 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.822387 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.822415 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.822434 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:44Z","lastTransitionTime":"2025-10-10T13:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.925496 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.925535 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.925546 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.925562 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:44 crc kubenswrapper[4745]: I1010 13:18:44.925583 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:44Z","lastTransitionTime":"2025-10-10T13:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.029096 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.029188 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.029215 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.029249 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.029275 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:45Z","lastTransitionTime":"2025-10-10T13:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.132276 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.132346 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.132363 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.132389 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.132407 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:45Z","lastTransitionTime":"2025-10-10T13:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.236156 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.236213 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.236229 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.236253 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.236270 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:45Z","lastTransitionTime":"2025-10-10T13:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.339594 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.339655 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.339672 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.339696 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.339712 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:45Z","lastTransitionTime":"2025-10-10T13:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.442981 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.443030 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.443047 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.443069 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.443086 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:45Z","lastTransitionTime":"2025-10-10T13:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.546124 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.546190 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.546208 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.546234 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.546252 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:45Z","lastTransitionTime":"2025-10-10T13:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.648868 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.648924 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.648945 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.648983 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.649004 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:45Z","lastTransitionTime":"2025-10-10T13:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.744118 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.744202 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:45 crc kubenswrapper[4745]: E1010 13:18:45.744318 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:18:45 crc kubenswrapper[4745]: E1010 13:18:45.744436 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.752306 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.752374 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.752392 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.752422 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.752446 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:45Z","lastTransitionTime":"2025-10-10T13:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.845541 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs\") pod \"network-metrics-daemon-xgh8p\" (UID: \"82c4e63a-009f-4720-87b5-640a8ec65117\") " pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:45 crc kubenswrapper[4745]: E1010 13:18:45.845810 4745 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 13:18:45 crc kubenswrapper[4745]: E1010 13:18:45.845953 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs podName:82c4e63a-009f-4720-87b5-640a8ec65117 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:49.845913145 +0000 UTC m=+43.743569948 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs") pod "network-metrics-daemon-xgh8p" (UID: "82c4e63a-009f-4720-87b5-640a8ec65117") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.855995 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.856224 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.856382 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.856519 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.856673 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:45Z","lastTransitionTime":"2025-10-10T13:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.960299 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.960602 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.960672 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.960768 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:45 crc kubenswrapper[4745]: I1010 13:18:45.960860 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:45Z","lastTransitionTime":"2025-10-10T13:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.065288 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.065349 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.065367 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.065395 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.065416 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:46Z","lastTransitionTime":"2025-10-10T13:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.168485 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.168541 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.168559 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.168587 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.168606 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:46Z","lastTransitionTime":"2025-10-10T13:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.271913 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.271958 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.271974 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.271997 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.272013 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:46Z","lastTransitionTime":"2025-10-10T13:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.375524 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.375603 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.375626 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.375657 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.375680 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:46Z","lastTransitionTime":"2025-10-10T13:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.479056 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.479130 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.479151 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.479184 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.479206 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:46Z","lastTransitionTime":"2025-10-10T13:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.581807 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.581857 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.581872 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.581894 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.581909 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:46Z","lastTransitionTime":"2025-10-10T13:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.684887 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.684933 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.684945 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.684964 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.684978 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:46Z","lastTransitionTime":"2025-10-10T13:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.744985 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.745086 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:46 crc kubenswrapper[4745]: E1010 13:18:46.745165 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:46 crc kubenswrapper[4745]: E1010 13:18:46.745371 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.772647 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:46Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.788361 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.788454 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.788503 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.788532 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.788551 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:46Z","lastTransitionTime":"2025-10-10T13:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.797946 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:46Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.819119 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:46Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.842982 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:46Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.862900 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:46Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.891624 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.891787 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.891816 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.891884 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.891912 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:46Z","lastTransitionTime":"2025-10-10T13:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.899285 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"message\\\":\\\"de network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z]\\\\nI1010 13:18:38.877288 6158 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-7rdsn after 0 failed attempt(s)\\\\nI1010 13:18:38.877300 6158 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7rdsn\\\\nI1010 13:18:38.876514 6158 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1010 13:18:38.876469 6158 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:46Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.952402 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:46Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.979415 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:46Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.994446 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.994512 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.994523 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.994544 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:46 crc kubenswrapper[4745]: I1010 13:18:46.994555 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:46Z","lastTransitionTime":"2025-10-10T13:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.001220 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:46Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.016155 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:47Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.032690 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:47Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.048405 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:47Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.064460 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:47Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.077765 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:47Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.091603 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:47Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.097312 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.097385 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.097411 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.097442 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.097466 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:47Z","lastTransitionTime":"2025-10-10T13:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.105828 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:47Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.122876 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82c4e63a-009f-4720-87b5-640a8ec65117\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xgh8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:47Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.201452 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.201519 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.201537 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.201565 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.201584 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:47Z","lastTransitionTime":"2025-10-10T13:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.304829 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.304909 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.304925 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.304950 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.304972 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:47Z","lastTransitionTime":"2025-10-10T13:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.407998 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.408043 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.408057 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.408078 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.408090 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:47Z","lastTransitionTime":"2025-10-10T13:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.511225 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.511370 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.511390 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.511414 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.511431 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:47Z","lastTransitionTime":"2025-10-10T13:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.615078 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.615151 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.615172 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.615199 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.615216 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:47Z","lastTransitionTime":"2025-10-10T13:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.719598 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.719667 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.719684 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.719711 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.719762 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:47Z","lastTransitionTime":"2025-10-10T13:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.744436 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:47 crc kubenswrapper[4745]: E1010 13:18:47.744666 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.744947 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:47 crc kubenswrapper[4745]: E1010 13:18:47.745326 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.823553 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.823618 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.823629 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.823649 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.823660 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:47Z","lastTransitionTime":"2025-10-10T13:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.926784 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.926858 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.926879 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.926904 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:47 crc kubenswrapper[4745]: I1010 13:18:47.926922 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:47Z","lastTransitionTime":"2025-10-10T13:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.030312 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.030385 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.030405 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.030437 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.030464 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:48Z","lastTransitionTime":"2025-10-10T13:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.134759 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.134828 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.134846 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.134873 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.134891 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:48Z","lastTransitionTime":"2025-10-10T13:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.238497 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.239481 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.239524 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.239554 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.239572 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:48Z","lastTransitionTime":"2025-10-10T13:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.342790 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.342856 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.342873 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.342898 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.342916 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:48Z","lastTransitionTime":"2025-10-10T13:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.445550 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.445627 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.445654 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.445680 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.445703 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:48Z","lastTransitionTime":"2025-10-10T13:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.549395 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.549435 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.549445 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.549461 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.549497 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:48Z","lastTransitionTime":"2025-10-10T13:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.652692 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.652743 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.652754 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.652786 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.652795 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:48Z","lastTransitionTime":"2025-10-10T13:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.744355 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.744499 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:48 crc kubenswrapper[4745]: E1010 13:18:48.744579 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:48 crc kubenswrapper[4745]: E1010 13:18:48.744918 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.755341 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.755381 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.755390 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.755403 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.755412 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:48Z","lastTransitionTime":"2025-10-10T13:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.859216 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.859414 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.859451 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.859528 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.859556 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:48Z","lastTransitionTime":"2025-10-10T13:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.963438 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.963519 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.963538 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.963566 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:48 crc kubenswrapper[4745]: I1010 13:18:48.963591 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:48Z","lastTransitionTime":"2025-10-10T13:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.066687 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.066750 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.066763 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.066787 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.066798 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:49Z","lastTransitionTime":"2025-10-10T13:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.170960 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.171046 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.171060 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.171082 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.171096 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:49Z","lastTransitionTime":"2025-10-10T13:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.274627 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.274712 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.274770 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.274811 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.274837 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:49Z","lastTransitionTime":"2025-10-10T13:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.378850 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.378926 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.378963 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.378997 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.379022 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:49Z","lastTransitionTime":"2025-10-10T13:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.482252 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.482327 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.482350 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.482384 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.482407 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:49Z","lastTransitionTime":"2025-10-10T13:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.586437 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.586483 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.586492 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.586512 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.586523 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:49Z","lastTransitionTime":"2025-10-10T13:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.690229 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.690471 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.690504 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.690536 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.690558 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:49Z","lastTransitionTime":"2025-10-10T13:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.744972 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.745078 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:49 crc kubenswrapper[4745]: E1010 13:18:49.745400 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:49 crc kubenswrapper[4745]: E1010 13:18:49.745594 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.793007 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.793376 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.793449 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.793574 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.793656 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:49Z","lastTransitionTime":"2025-10-10T13:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.895369 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs\") pod \"network-metrics-daemon-xgh8p\" (UID: \"82c4e63a-009f-4720-87b5-640a8ec65117\") " pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:49 crc kubenswrapper[4745]: E1010 13:18:49.896228 4745 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 13:18:49 crc kubenswrapper[4745]: E1010 13:18:49.896721 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs podName:82c4e63a-009f-4720-87b5-640a8ec65117 nodeName:}" failed. No retries permitted until 2025-10-10 13:18:57.89652712 +0000 UTC m=+51.794184053 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs") pod "network-metrics-daemon-xgh8p" (UID: "82c4e63a-009f-4720-87b5-640a8ec65117") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.897090 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.897389 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.897553 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.897708 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:49 crc kubenswrapper[4745]: I1010 13:18:49.897944 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:49Z","lastTransitionTime":"2025-10-10T13:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.001103 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.001185 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.001210 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.001244 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.001261 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:50Z","lastTransitionTime":"2025-10-10T13:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.104243 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.104315 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.104327 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.104352 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.104368 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:50Z","lastTransitionTime":"2025-10-10T13:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.207822 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.207870 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.207884 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.207906 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.207920 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:50Z","lastTransitionTime":"2025-10-10T13:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.311671 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.311786 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.311806 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.311836 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.311854 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:50Z","lastTransitionTime":"2025-10-10T13:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.340449 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.342294 4745 scope.go:117] "RemoveContainer" containerID="c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.416152 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.416197 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.416208 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.416224 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.416235 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:50Z","lastTransitionTime":"2025-10-10T13:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.519625 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.520128 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.520139 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.520160 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.520173 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:50Z","lastTransitionTime":"2025-10-10T13:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.623527 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.623583 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.623600 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.623627 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.623645 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:50Z","lastTransitionTime":"2025-10-10T13:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.726387 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.726456 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.726474 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.726506 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.726526 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:50Z","lastTransitionTime":"2025-10-10T13:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.744775 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:50 crc kubenswrapper[4745]: E1010 13:18:50.744949 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.745057 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:50 crc kubenswrapper[4745]: E1010 13:18:50.745351 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.830385 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.830454 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.830471 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.830496 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.830513 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:50Z","lastTransitionTime":"2025-10-10T13:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.933817 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.933871 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.933883 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.933902 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:50 crc kubenswrapper[4745]: I1010 13:18:50.933914 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:50Z","lastTransitionTime":"2025-10-10T13:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.036379 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.036460 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.036482 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.036513 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.036531 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:51Z","lastTransitionTime":"2025-10-10T13:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.080482 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovnkube-controller/1.log" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.084534 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerStarted","Data":"7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e"} Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.085404 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.108517 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:51Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.139442 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.139503 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.139522 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.139550 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.139570 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:51Z","lastTransitionTime":"2025-10-10T13:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.151300 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:51Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.184978 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:51Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.201011 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:51Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.218656 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:51Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.236431 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:51Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.242483 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.242527 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.242538 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.242554 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.242564 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:51Z","lastTransitionTime":"2025-10-10T13:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.269338 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"message\\\":\\\"de network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z]\\\\nI1010 13:18:38.877288 6158 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-7rdsn after 0 failed attempt(s)\\\\nI1010 13:18:38.877300 6158 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7rdsn\\\\nI1010 13:18:38.876514 6158 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1010 13:18:38.876469 6158 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:51Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.288938 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:51Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.309618 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:51Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.322783 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:51Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.336649 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:51Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.345712 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.345773 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.345786 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.345811 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.345822 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:51Z","lastTransitionTime":"2025-10-10T13:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.357418 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:51Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.379045 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:51Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.392123 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:51Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.407031 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:51Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.418666 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:51Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.432401 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82c4e63a-009f-4720-87b5-640a8ec65117\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xgh8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:51Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.448431 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.448655 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.448846 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.449050 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.449211 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:51Z","lastTransitionTime":"2025-10-10T13:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.552550 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.552618 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.552635 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.552663 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.552683 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:51Z","lastTransitionTime":"2025-10-10T13:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.655915 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.655972 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.655992 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.656020 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.656041 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:51Z","lastTransitionTime":"2025-10-10T13:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.744974 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.745046 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:51 crc kubenswrapper[4745]: E1010 13:18:51.745155 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:51 crc kubenswrapper[4745]: E1010 13:18:51.745325 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.759139 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.759213 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.759231 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.759259 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.759281 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:51Z","lastTransitionTime":"2025-10-10T13:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.862462 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.862524 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.862542 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.862572 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.862594 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:51Z","lastTransitionTime":"2025-10-10T13:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.965923 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.966010 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.966029 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.966063 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:51 crc kubenswrapper[4745]: I1010 13:18:51.966085 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:51Z","lastTransitionTime":"2025-10-10T13:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.069694 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.069773 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.069793 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.069818 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.069839 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:52Z","lastTransitionTime":"2025-10-10T13:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.091839 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovnkube-controller/2.log" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.093091 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovnkube-controller/1.log" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.098531 4745 generic.go:334] "Generic (PLEG): container finished" podID="687bad4e-baff-4e09-bb33-def12c499092" containerID="7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e" exitCode=1 Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.098626 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerDied","Data":"7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e"} Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.098765 4745 scope.go:117] "RemoveContainer" containerID="c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.103408 4745 scope.go:117] "RemoveContainer" containerID="7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e" Oct 10 13:18:52 crc kubenswrapper[4745]: E1010 13:18:52.103966 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podUID="687bad4e-baff-4e09-bb33-def12c499092" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.121580 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:52Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.142399 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:52Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.159839 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:52Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.172388 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.172592 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.172649 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.172711 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.172827 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:52Z","lastTransitionTime":"2025-10-10T13:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.174033 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:52Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.191410 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:52Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.206209 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:52Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.217531 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82c4e63a-009f-4720-87b5-640a8ec65117\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xgh8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:52Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.229050 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:52Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.258218 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c54b29904d12c4553cc0eb55d67b2ca6f308c761165d121426ceaed0c3d5bc52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"message\\\":\\\"de network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:38Z is after 2025-08-24T17:21:41Z]\\\\nI1010 13:18:38.877288 6158 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-7rdsn after 0 failed attempt(s)\\\\nI1010 13:18:38.877300 6158 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7rdsn\\\\nI1010 13:18:38.876514 6158 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1010 13:18:38.876469 6158 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:51Z\\\",\\\"message\\\":\\\"6 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 13:18:51.354210 6376 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 13:18:51.354660 6376 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354695 6376 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354832 6376 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355149 6376 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355386 6376 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1010 13:18:51.355631 6376 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:52Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.276223 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.276323 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.276352 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.276390 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.276418 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:52Z","lastTransitionTime":"2025-10-10T13:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.277807 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:52Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.299358 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:52Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.314469 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:52Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.329329 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:52Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.346690 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:52Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.362631 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:52Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.377382 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:52Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.379550 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.379611 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.379629 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.379657 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.379678 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:52Z","lastTransitionTime":"2025-10-10T13:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.390309 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:52Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.482977 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.483042 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.483058 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.483079 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.483094 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:52Z","lastTransitionTime":"2025-10-10T13:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.586088 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.586158 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.586182 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.586216 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.586242 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:52Z","lastTransitionTime":"2025-10-10T13:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.689235 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.689311 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.689333 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.689409 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.689503 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:52Z","lastTransitionTime":"2025-10-10T13:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.744255 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.744363 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:52 crc kubenswrapper[4745]: E1010 13:18:52.744409 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:52 crc kubenswrapper[4745]: E1010 13:18:52.744635 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.792236 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.792312 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.792331 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.792366 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.792391 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:52Z","lastTransitionTime":"2025-10-10T13:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.896662 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.896776 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.896800 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.896825 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:52 crc kubenswrapper[4745]: I1010 13:18:52.896881 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:52Z","lastTransitionTime":"2025-10-10T13:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.000494 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.000662 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.000685 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.000712 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.000793 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:53Z","lastTransitionTime":"2025-10-10T13:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.103623 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.103704 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.103719 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.103780 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.103812 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:53Z","lastTransitionTime":"2025-10-10T13:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.105249 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovnkube-controller/2.log" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.109633 4745 scope.go:117] "RemoveContainer" containerID="7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e" Oct 10 13:18:53 crc kubenswrapper[4745]: E1010 13:18:53.109831 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podUID="687bad4e-baff-4e09-bb33-def12c499092" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.127645 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.146816 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.168322 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.184499 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82c4e63a-009f-4720-87b5-640a8ec65117\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xgh8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.207500 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.207606 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.207627 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.207688 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.207709 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:53Z","lastTransitionTime":"2025-10-10T13:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.216051 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.239129 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.261872 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.286971 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.308536 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.310863 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.310938 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.310952 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.310970 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.310982 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:53Z","lastTransitionTime":"2025-10-10T13:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.337926 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:51Z\\\",\\\"message\\\":\\\"6 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 13:18:51.354210 6376 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 13:18:51.354660 6376 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354695 6376 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354832 6376 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355149 6376 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355386 6376 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1010 13:18:51.355631 6376 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.365548 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.388372 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.409452 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.414241 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.414303 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.414316 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.414336 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.414348 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:53Z","lastTransitionTime":"2025-10-10T13:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.427016 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.446286 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.470470 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.492964 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.518157 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.518209 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.518226 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.518253 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.518273 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:53Z","lastTransitionTime":"2025-10-10T13:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.610798 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.611261 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.611546 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.612086 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.612421 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:53Z","lastTransitionTime":"2025-10-10T13:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:53 crc kubenswrapper[4745]: E1010 13:18:53.632620 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.639084 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.639160 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.639178 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.639210 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.639225 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:53Z","lastTransitionTime":"2025-10-10T13:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:53 crc kubenswrapper[4745]: E1010 13:18:53.660335 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.666847 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.666914 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.666937 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.666972 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.666997 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:53Z","lastTransitionTime":"2025-10-10T13:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:53 crc kubenswrapper[4745]: E1010 13:18:53.689120 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.695239 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.695335 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.695354 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.695392 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.695416 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:53Z","lastTransitionTime":"2025-10-10T13:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:53 crc kubenswrapper[4745]: E1010 13:18:53.718320 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.723980 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.724033 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.724047 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.724073 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.724091 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:53Z","lastTransitionTime":"2025-10-10T13:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.744391 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:53 crc kubenswrapper[4745]: E1010 13:18:53.744590 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.744687 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:53 crc kubenswrapper[4745]: E1010 13:18:53.745099 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:53 crc kubenswrapper[4745]: E1010 13:18:53.745070 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:53Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:53 crc kubenswrapper[4745]: E1010 13:18:53.745234 4745 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.750150 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.750191 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.750205 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.750227 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.750242 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:53Z","lastTransitionTime":"2025-10-10T13:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.854943 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.855033 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.855046 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.855074 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.855087 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:53Z","lastTransitionTime":"2025-10-10T13:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.958949 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.959023 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.959041 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.959070 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:53 crc kubenswrapper[4745]: I1010 13:18:53.959091 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:53Z","lastTransitionTime":"2025-10-10T13:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.062768 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.062864 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.062892 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.062929 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.062959 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:54Z","lastTransitionTime":"2025-10-10T13:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.166491 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.166572 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.166592 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.166623 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.166643 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:54Z","lastTransitionTime":"2025-10-10T13:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.270679 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.270776 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.270793 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.270819 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.270839 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:54Z","lastTransitionTime":"2025-10-10T13:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.374280 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.374337 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.374348 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.374370 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.374386 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:54Z","lastTransitionTime":"2025-10-10T13:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.478353 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.478420 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.478438 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.478468 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.478487 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:54Z","lastTransitionTime":"2025-10-10T13:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.581457 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.581504 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.581518 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.581535 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.581547 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:54Z","lastTransitionTime":"2025-10-10T13:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.685042 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.685146 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.685164 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.685187 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.685204 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:54Z","lastTransitionTime":"2025-10-10T13:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.744468 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.744579 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:54 crc kubenswrapper[4745]: E1010 13:18:54.744663 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:54 crc kubenswrapper[4745]: E1010 13:18:54.744813 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.788946 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.789027 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.789048 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.789080 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.789101 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:54Z","lastTransitionTime":"2025-10-10T13:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.892522 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.892564 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.892575 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.892594 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.892607 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:54Z","lastTransitionTime":"2025-10-10T13:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.995965 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.996019 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.996032 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.996054 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:54 crc kubenswrapper[4745]: I1010 13:18:54.996068 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:54Z","lastTransitionTime":"2025-10-10T13:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.100227 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.100696 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.100877 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.101113 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.101246 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:55Z","lastTransitionTime":"2025-10-10T13:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.206033 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.206103 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.206113 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.206132 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.206144 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:55Z","lastTransitionTime":"2025-10-10T13:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.308996 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.309047 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.309058 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.309080 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.309094 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:55Z","lastTransitionTime":"2025-10-10T13:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.412482 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.412558 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.412578 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.412614 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.412636 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:55Z","lastTransitionTime":"2025-10-10T13:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.516402 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.516485 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.516506 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.516536 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.516559 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:55Z","lastTransitionTime":"2025-10-10T13:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.620251 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.620346 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.620369 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.620407 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.620428 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:55Z","lastTransitionTime":"2025-10-10T13:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.723971 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.724066 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.724090 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.724122 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.724142 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:55Z","lastTransitionTime":"2025-10-10T13:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.744352 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.744391 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:55 crc kubenswrapper[4745]: E1010 13:18:55.744597 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:18:55 crc kubenswrapper[4745]: E1010 13:18:55.744802 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.827979 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.828062 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.828087 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.828121 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.828145 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:55Z","lastTransitionTime":"2025-10-10T13:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.932317 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.932397 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.932415 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.932449 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:55 crc kubenswrapper[4745]: I1010 13:18:55.932468 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:55Z","lastTransitionTime":"2025-10-10T13:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.036440 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.036518 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.036537 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.036568 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.036612 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:56Z","lastTransitionTime":"2025-10-10T13:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.140061 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.140692 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.140942 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.141156 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.141300 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:56Z","lastTransitionTime":"2025-10-10T13:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.244701 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.244785 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.244799 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.244837 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.244853 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:56Z","lastTransitionTime":"2025-10-10T13:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.348679 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.348784 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.348805 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.348833 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.348852 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:56Z","lastTransitionTime":"2025-10-10T13:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.452186 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.452237 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.452251 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.452273 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.452287 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:56Z","lastTransitionTime":"2025-10-10T13:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.555345 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.555702 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.555869 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.556005 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.556159 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:56Z","lastTransitionTime":"2025-10-10T13:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.659713 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.660148 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.660304 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.660448 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.660611 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:56Z","lastTransitionTime":"2025-10-10T13:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.744583 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.744583 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:56 crc kubenswrapper[4745]: E1010 13:18:56.744887 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:56 crc kubenswrapper[4745]: E1010 13:18:56.745279 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.763412 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.763499 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.763520 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.763552 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.763574 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:56Z","lastTransitionTime":"2025-10-10T13:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.769828 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:56Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.795780 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:56Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.809073 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:56Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.825154 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:56Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.841679 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:56Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.866282 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.866398 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.866421 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.866451 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.866473 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:56Z","lastTransitionTime":"2025-10-10T13:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.872110 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:56Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.890484 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:56Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.908218 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:56Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.924164 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:56Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.943867 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82c4e63a-009f-4720-87b5-640a8ec65117\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xgh8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:56Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.963972 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:56Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.969146 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.969250 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.969265 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.969291 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.969308 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:56Z","lastTransitionTime":"2025-10-10T13:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:56 crc kubenswrapper[4745]: I1010 13:18:56.991853 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:56Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.015097 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:57Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.037329 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:57Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.062786 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:57Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.071592 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.071646 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.071662 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.071685 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.071704 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:57Z","lastTransitionTime":"2025-10-10T13:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.077113 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:57Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.104921 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:51Z\\\",\\\"message\\\":\\\"6 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 13:18:51.354210 6376 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 13:18:51.354660 6376 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354695 6376 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354832 6376 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355149 6376 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355386 6376 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1010 13:18:51.355631 6376 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:57Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.175590 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.175681 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.175699 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.175767 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.175788 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:57Z","lastTransitionTime":"2025-10-10T13:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.279780 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.279934 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.280028 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.280078 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.280106 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:57Z","lastTransitionTime":"2025-10-10T13:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.384129 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.384220 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.384246 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.384282 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.384313 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:57Z","lastTransitionTime":"2025-10-10T13:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.487996 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.488039 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.488049 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.488068 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.488078 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:57Z","lastTransitionTime":"2025-10-10T13:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.591376 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.591439 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.591453 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.591471 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.591504 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:57Z","lastTransitionTime":"2025-10-10T13:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.694404 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.694465 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.694477 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.694511 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.694521 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:57Z","lastTransitionTime":"2025-10-10T13:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.744429 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.744488 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:57 crc kubenswrapper[4745]: E1010 13:18:57.745024 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:18:57 crc kubenswrapper[4745]: E1010 13:18:57.745300 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.797437 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.797492 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.797510 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.797535 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.797552 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:57Z","lastTransitionTime":"2025-10-10T13:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.900383 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.900473 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.900498 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.900531 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.900555 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:57Z","lastTransitionTime":"2025-10-10T13:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:57 crc kubenswrapper[4745]: I1010 13:18:57.908873 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs\") pod \"network-metrics-daemon-xgh8p\" (UID: \"82c4e63a-009f-4720-87b5-640a8ec65117\") " pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:57 crc kubenswrapper[4745]: E1010 13:18:57.909017 4745 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 13:18:57 crc kubenswrapper[4745]: E1010 13:18:57.909085 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs podName:82c4e63a-009f-4720-87b5-640a8ec65117 nodeName:}" failed. No retries permitted until 2025-10-10 13:19:13.909067805 +0000 UTC m=+67.806724578 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs") pod "network-metrics-daemon-xgh8p" (UID: "82c4e63a-009f-4720-87b5-640a8ec65117") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.004410 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.004442 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.004459 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.004479 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.004492 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:58Z","lastTransitionTime":"2025-10-10T13:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.009962 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:18:58 crc kubenswrapper[4745]: E1010 13:18:58.010147 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:19:30.010133791 +0000 UTC m=+83.907790564 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.108331 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.108595 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.108690 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.108796 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.108974 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:58Z","lastTransitionTime":"2025-10-10T13:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.111032 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.111110 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.111170 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.111218 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:58 crc kubenswrapper[4745]: E1010 13:18:58.111409 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 13:18:58 crc kubenswrapper[4745]: E1010 13:18:58.111515 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 13:18:58 crc kubenswrapper[4745]: E1010 13:18:58.111571 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 13:18:58 crc kubenswrapper[4745]: E1010 13:18:58.111534 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 13:18:58 crc kubenswrapper[4745]: E1010 13:18:58.111619 4745 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:58 crc kubenswrapper[4745]: E1010 13:18:58.111697 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 13:19:30.111672309 +0000 UTC m=+84.009329142 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:58 crc kubenswrapper[4745]: E1010 13:18:58.111457 4745 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 13:18:58 crc kubenswrapper[4745]: E1010 13:18:58.111934 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 13:19:30.111918705 +0000 UTC m=+84.009575498 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 13:18:58 crc kubenswrapper[4745]: E1010 13:18:58.111410 4745 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 13:18:58 crc kubenswrapper[4745]: E1010 13:18:58.112113 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 13:19:30.112102219 +0000 UTC m=+84.009759012 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 13:18:58 crc kubenswrapper[4745]: E1010 13:18:58.111588 4745 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:58 crc kubenswrapper[4745]: E1010 13:18:58.112293 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 13:19:30.112283264 +0000 UTC m=+84.009940037 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.212220 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.212267 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.212279 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.212296 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.212308 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:58Z","lastTransitionTime":"2025-10-10T13:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.316178 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.316259 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.316295 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.316324 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.316345 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:58Z","lastTransitionTime":"2025-10-10T13:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.401929 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.420288 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.420631 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.420864 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.420502 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.421252 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.421419 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:58Z","lastTransitionTime":"2025-10-10T13:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.424841 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:58Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.443605 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:58Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.466723 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:58Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.491304 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:58Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.514084 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:58Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.525210 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.525445 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.525575 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.525699 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.525869 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:58Z","lastTransitionTime":"2025-10-10T13:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.536992 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:58Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.561055 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:58Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.577270 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:58Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.588865 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82c4e63a-009f-4720-87b5-640a8ec65117\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xgh8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:58Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.604327 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:58Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.629248 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.629318 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.629337 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.629367 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.629388 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:58Z","lastTransitionTime":"2025-10-10T13:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.630234 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:58Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.650844 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:58Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.677184 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:58Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.694403 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:58Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.726906 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:51Z\\\",\\\"message\\\":\\\"6 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 13:18:51.354210 6376 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 13:18:51.354660 6376 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354695 6376 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354832 6376 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355149 6376 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355386 6376 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1010 13:18:51.355631 6376 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:58Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.731724 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.731903 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.731924 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.731986 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.732008 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:58Z","lastTransitionTime":"2025-10-10T13:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.744428 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:18:58 crc kubenswrapper[4745]: E1010 13:18:58.744529 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.744629 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:18:58 crc kubenswrapper[4745]: E1010 13:18:58.744854 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.752247 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:58Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.784414 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:18:58Z is after 2025-08-24T17:21:41Z" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.836302 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.836393 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.836415 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.836447 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.836468 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:58Z","lastTransitionTime":"2025-10-10T13:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.939858 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.939932 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.939954 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.939989 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:58 crc kubenswrapper[4745]: I1010 13:18:58.940015 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:58Z","lastTransitionTime":"2025-10-10T13:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.043854 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.043925 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.043946 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.043974 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.043995 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:59Z","lastTransitionTime":"2025-10-10T13:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.147200 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.147267 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.147286 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.147315 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.147333 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:59Z","lastTransitionTime":"2025-10-10T13:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.250707 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.250798 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.250816 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.250839 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.250856 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:59Z","lastTransitionTime":"2025-10-10T13:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.353906 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.353972 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.353990 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.354015 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.354035 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:59Z","lastTransitionTime":"2025-10-10T13:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.457029 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.457101 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.457123 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.457149 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.457167 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:59Z","lastTransitionTime":"2025-10-10T13:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.560679 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.560762 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.560777 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.560795 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.560807 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:59Z","lastTransitionTime":"2025-10-10T13:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.664916 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.664959 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.664970 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.664986 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.665000 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:59Z","lastTransitionTime":"2025-10-10T13:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.744675 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.744719 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:18:59 crc kubenswrapper[4745]: E1010 13:18:59.744916 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:18:59 crc kubenswrapper[4745]: E1010 13:18:59.745142 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.769002 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.769052 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.769070 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.769093 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.769112 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:59Z","lastTransitionTime":"2025-10-10T13:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.873050 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.873108 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.873125 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.873149 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.873168 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:59Z","lastTransitionTime":"2025-10-10T13:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.976215 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.976283 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.976306 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.976335 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:18:59 crc kubenswrapper[4745]: I1010 13:18:59.976358 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:18:59Z","lastTransitionTime":"2025-10-10T13:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.079799 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.079912 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.079937 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.079967 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.079989 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:00Z","lastTransitionTime":"2025-10-10T13:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.183690 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.183801 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.183819 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.184215 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.184263 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:00Z","lastTransitionTime":"2025-10-10T13:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.287318 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.287390 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.287412 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.287441 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.287467 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:00Z","lastTransitionTime":"2025-10-10T13:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.390892 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.390956 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.390972 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.390995 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.391013 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:00Z","lastTransitionTime":"2025-10-10T13:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.493978 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.494046 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.494066 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.494092 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.494110 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:00Z","lastTransitionTime":"2025-10-10T13:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.597268 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.597342 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.597368 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.597397 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.597420 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:00Z","lastTransitionTime":"2025-10-10T13:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.700357 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.700416 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.700438 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.700466 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.700487 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:00Z","lastTransitionTime":"2025-10-10T13:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.744071 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:00 crc kubenswrapper[4745]: E1010 13:19:00.744216 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.744419 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:00 crc kubenswrapper[4745]: E1010 13:19:00.744524 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.803498 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.803877 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.804026 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.804166 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.804302 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:00Z","lastTransitionTime":"2025-10-10T13:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.907384 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.907465 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.907489 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.907524 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:00 crc kubenswrapper[4745]: I1010 13:19:00.907547 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:00Z","lastTransitionTime":"2025-10-10T13:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.010985 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.011097 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.011121 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.011149 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.011171 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:01Z","lastTransitionTime":"2025-10-10T13:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.114401 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.114490 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.114515 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.114541 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.114560 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:01Z","lastTransitionTime":"2025-10-10T13:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.218073 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.218124 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.218140 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.218162 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.218179 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:01Z","lastTransitionTime":"2025-10-10T13:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.320906 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.320957 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.320973 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.320991 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.321002 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:01Z","lastTransitionTime":"2025-10-10T13:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.424157 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.424217 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.424235 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.424260 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.424277 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:01Z","lastTransitionTime":"2025-10-10T13:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.527084 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.527144 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.527161 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.527213 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.527234 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:01Z","lastTransitionTime":"2025-10-10T13:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.630177 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.630245 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.630265 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.630293 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.630316 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:01Z","lastTransitionTime":"2025-10-10T13:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.733268 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.733325 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.733349 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.733373 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.733391 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:01Z","lastTransitionTime":"2025-10-10T13:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.744918 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.744942 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:01 crc kubenswrapper[4745]: E1010 13:19:01.745110 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:01 crc kubenswrapper[4745]: E1010 13:19:01.745262 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.837028 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.837114 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.837148 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.837180 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.837203 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:01Z","lastTransitionTime":"2025-10-10T13:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.940456 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.940537 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.940560 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.940591 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:01 crc kubenswrapper[4745]: I1010 13:19:01.940616 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:01Z","lastTransitionTime":"2025-10-10T13:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.043680 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.043818 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.043844 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.043871 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.043904 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:02Z","lastTransitionTime":"2025-10-10T13:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.146270 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.146335 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.146351 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.146376 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.146399 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:02Z","lastTransitionTime":"2025-10-10T13:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.250039 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.250109 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.250126 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.250151 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.250174 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:02Z","lastTransitionTime":"2025-10-10T13:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.353387 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.353456 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.353466 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.353482 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.353491 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:02Z","lastTransitionTime":"2025-10-10T13:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.456408 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.456461 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.456469 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.456482 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.456493 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:02Z","lastTransitionTime":"2025-10-10T13:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.558924 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.558986 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.559006 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.559057 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.559077 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:02Z","lastTransitionTime":"2025-10-10T13:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.663152 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.663219 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.663244 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.663277 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.663298 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:02Z","lastTransitionTime":"2025-10-10T13:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.744055 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:02 crc kubenswrapper[4745]: E1010 13:19:02.744241 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.744379 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:02 crc kubenswrapper[4745]: E1010 13:19:02.744602 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.765436 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.765493 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.765510 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.765532 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.765553 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:02Z","lastTransitionTime":"2025-10-10T13:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.868468 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.868574 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.868592 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.868616 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.868635 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:02Z","lastTransitionTime":"2025-10-10T13:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.972439 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.972502 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.972518 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.972549 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:02 crc kubenswrapper[4745]: I1010 13:19:02.972571 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:02Z","lastTransitionTime":"2025-10-10T13:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.075821 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.075886 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.075902 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.075925 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.075942 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:03Z","lastTransitionTime":"2025-10-10T13:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.179031 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.179104 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.179129 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.179157 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.179179 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:03Z","lastTransitionTime":"2025-10-10T13:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.281968 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.282032 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.282051 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.282075 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.282094 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:03Z","lastTransitionTime":"2025-10-10T13:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.385068 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.385135 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.385154 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.385180 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.385199 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:03Z","lastTransitionTime":"2025-10-10T13:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.487348 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.487408 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.487431 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.487459 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.487479 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:03Z","lastTransitionTime":"2025-10-10T13:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.590537 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.590608 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.590624 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.590649 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.590672 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:03Z","lastTransitionTime":"2025-10-10T13:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.693994 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.694048 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.694066 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.694090 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.694107 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:03Z","lastTransitionTime":"2025-10-10T13:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.744293 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.744340 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:03 crc kubenswrapper[4745]: E1010 13:19:03.744489 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:03 crc kubenswrapper[4745]: E1010 13:19:03.744940 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.796837 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.796897 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.796916 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.796939 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.796955 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:03Z","lastTransitionTime":"2025-10-10T13:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.900207 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.900277 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.900301 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.900329 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.900351 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:03Z","lastTransitionTime":"2025-10-10T13:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.941110 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.941174 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.941191 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.941214 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.941231 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:03Z","lastTransitionTime":"2025-10-10T13:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:03 crc kubenswrapper[4745]: E1010 13:19:03.959028 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:03Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.964424 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.964492 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.964509 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.964532 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.964551 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:03Z","lastTransitionTime":"2025-10-10T13:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:03 crc kubenswrapper[4745]: E1010 13:19:03.984593 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:03Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.989340 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.989403 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.989424 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.989450 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:03 crc kubenswrapper[4745]: I1010 13:19:03.989468 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:03Z","lastTransitionTime":"2025-10-10T13:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:04 crc kubenswrapper[4745]: E1010 13:19:04.008676 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:04Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.013045 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.013113 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.013126 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.013146 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.013179 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:04Z","lastTransitionTime":"2025-10-10T13:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:04 crc kubenswrapper[4745]: E1010 13:19:04.032925 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:04Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.037840 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.037933 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.037948 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.037967 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.037999 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:04Z","lastTransitionTime":"2025-10-10T13:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:04 crc kubenswrapper[4745]: E1010 13:19:04.058882 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:04Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:04 crc kubenswrapper[4745]: E1010 13:19:04.059122 4745 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.064167 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.064351 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.064433 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.064476 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.064518 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:04Z","lastTransitionTime":"2025-10-10T13:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.167898 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.167963 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.168249 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.168592 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.168665 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:04Z","lastTransitionTime":"2025-10-10T13:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.272389 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.272434 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.272452 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.272475 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.272492 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:04Z","lastTransitionTime":"2025-10-10T13:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.374974 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.375061 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.375083 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.375514 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.375830 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:04Z","lastTransitionTime":"2025-10-10T13:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.479319 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.479379 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.479397 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.479435 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.479452 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:04Z","lastTransitionTime":"2025-10-10T13:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.582096 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.582153 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.582171 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.582194 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.582217 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:04Z","lastTransitionTime":"2025-10-10T13:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.685380 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.685449 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.685475 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.685504 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.685521 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:04Z","lastTransitionTime":"2025-10-10T13:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.744654 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.744684 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:04 crc kubenswrapper[4745]: E1010 13:19:04.744956 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:04 crc kubenswrapper[4745]: E1010 13:19:04.745070 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.788770 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.788853 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.788872 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.788899 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.788919 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:04Z","lastTransitionTime":"2025-10-10T13:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.892271 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.892333 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.892351 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.892374 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.892390 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:04Z","lastTransitionTime":"2025-10-10T13:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.996486 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.996605 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.996624 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.996648 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:04 crc kubenswrapper[4745]: I1010 13:19:04.996667 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:04Z","lastTransitionTime":"2025-10-10T13:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.099842 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.099902 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.099919 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.099947 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.099968 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:05Z","lastTransitionTime":"2025-10-10T13:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.203346 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.203397 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.203413 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.203435 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.203451 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:05Z","lastTransitionTime":"2025-10-10T13:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.307379 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.307444 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.307468 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.307499 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.307519 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:05Z","lastTransitionTime":"2025-10-10T13:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.411010 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.411063 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.411076 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.411094 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.411106 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:05Z","lastTransitionTime":"2025-10-10T13:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.514433 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.514504 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.514522 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.514545 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.514562 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:05Z","lastTransitionTime":"2025-10-10T13:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.617808 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.617869 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.617886 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.617911 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.617929 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:05Z","lastTransitionTime":"2025-10-10T13:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.727182 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.727240 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.727258 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.727465 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.727482 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:05Z","lastTransitionTime":"2025-10-10T13:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.744520 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.744529 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:05 crc kubenswrapper[4745]: E1010 13:19:05.744700 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:05 crc kubenswrapper[4745]: E1010 13:19:05.745424 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.745945 4745 scope.go:117] "RemoveContainer" containerID="7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e" Oct 10 13:19:05 crc kubenswrapper[4745]: E1010 13:19:05.746407 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podUID="687bad4e-baff-4e09-bb33-def12c499092" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.831205 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.831276 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.831298 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.831329 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.831351 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:05Z","lastTransitionTime":"2025-10-10T13:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.934676 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.934777 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.934795 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.934819 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:05 crc kubenswrapper[4745]: I1010 13:19:05.934839 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:05Z","lastTransitionTime":"2025-10-10T13:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.038096 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.038177 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.038199 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.038234 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.038258 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:06Z","lastTransitionTime":"2025-10-10T13:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.141240 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.141308 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.141333 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.141381 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.141409 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:06Z","lastTransitionTime":"2025-10-10T13:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.243868 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.243929 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.243948 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.243972 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.243991 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:06Z","lastTransitionTime":"2025-10-10T13:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.346382 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.346421 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.346429 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.346441 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.346451 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:06Z","lastTransitionTime":"2025-10-10T13:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.449642 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.450095 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.450115 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.450152 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.450170 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:06Z","lastTransitionTime":"2025-10-10T13:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.554587 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.554685 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.554711 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.554794 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.554824 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:06Z","lastTransitionTime":"2025-10-10T13:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.657859 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.658009 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.658032 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.658065 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.658095 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:06Z","lastTransitionTime":"2025-10-10T13:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.745063 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:06 crc kubenswrapper[4745]: E1010 13:19:06.745264 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.746617 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:06 crc kubenswrapper[4745]: E1010 13:19:06.747087 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.762296 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.762371 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.762389 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.762444 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.762464 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:06Z","lastTransitionTime":"2025-10-10T13:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.767257 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923c385c-8102-4892-a4a3-67b6408545ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://193833f7e032a32aae576702c6965f7ea2869114b86d1a0349e7ad787869c74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab4465f351419e32376cc0b2062225c6faac1e27dce51bcf3c1ca4bb590fc4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f6659c3fefc2b4892b34a407a26418aa86c780a7e03c9120794f4396c08913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:06Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.787944 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:06Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.810397 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:06Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.830970 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:06Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.848610 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82c4e63a-009f-4720-87b5-640a8ec65117\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xgh8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:06Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.864394 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:06Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.865329 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.865379 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.865397 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.865424 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.865443 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:06Z","lastTransitionTime":"2025-10-10T13:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.901348 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:51Z\\\",\\\"message\\\":\\\"6 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 13:18:51.354210 6376 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 13:18:51.354660 6376 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354695 6376 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354832 6376 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355149 6376 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355386 6376 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1010 13:18:51.355631 6376 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:06Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.924818 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:06Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.959864 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:06Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.968427 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.968517 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.968543 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.968583 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.968615 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:06Z","lastTransitionTime":"2025-10-10T13:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:06 crc kubenswrapper[4745]: I1010 13:19:06.982848 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:06Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.003211 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:07Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.025506 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:07Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.046912 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:07Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.071424 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:07Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.072904 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.073174 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.073573 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.073979 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.074325 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:07Z","lastTransitionTime":"2025-10-10T13:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.090968 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:07Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.116439 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:07Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.144293 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:07Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.172344 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:07Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.178044 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.178140 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.178172 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.178211 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.178241 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:07Z","lastTransitionTime":"2025-10-10T13:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.280937 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.281047 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.281074 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.281118 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.281145 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:07Z","lastTransitionTime":"2025-10-10T13:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.384878 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.385338 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.385485 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.385629 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.385801 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:07Z","lastTransitionTime":"2025-10-10T13:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.489167 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.489240 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.489266 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.489307 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.489332 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:07Z","lastTransitionTime":"2025-10-10T13:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.593660 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.594102 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.594263 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.594435 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.594595 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:07Z","lastTransitionTime":"2025-10-10T13:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.698634 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.698683 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.698699 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.698722 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.698754 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:07Z","lastTransitionTime":"2025-10-10T13:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.744617 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.744830 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:07 crc kubenswrapper[4745]: E1010 13:19:07.744931 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:07 crc kubenswrapper[4745]: E1010 13:19:07.745063 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.802194 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.802256 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.802278 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.802309 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.802331 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:07Z","lastTransitionTime":"2025-10-10T13:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.906026 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.906598 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.906893 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.907126 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:07 crc kubenswrapper[4745]: I1010 13:19:07.907377 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:07Z","lastTransitionTime":"2025-10-10T13:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.011196 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.011879 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.012103 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.012368 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.012589 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:08Z","lastTransitionTime":"2025-10-10T13:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.117004 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.117077 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.117096 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.117124 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.117144 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:08Z","lastTransitionTime":"2025-10-10T13:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.219968 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.220036 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.220050 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.220073 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.220089 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:08Z","lastTransitionTime":"2025-10-10T13:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.323218 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.323296 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.323315 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.323342 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.323361 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:08Z","lastTransitionTime":"2025-10-10T13:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.427227 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.427313 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.427337 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.427368 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.427386 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:08Z","lastTransitionTime":"2025-10-10T13:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.531312 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.531432 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.531452 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.531476 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.531494 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:08Z","lastTransitionTime":"2025-10-10T13:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.635425 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.635459 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.635468 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.635481 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.635490 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:08Z","lastTransitionTime":"2025-10-10T13:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.737865 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.737904 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.737914 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.737929 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.737940 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:08Z","lastTransitionTime":"2025-10-10T13:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.744389 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.744517 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:08 crc kubenswrapper[4745]: E1010 13:19:08.744669 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:08 crc kubenswrapper[4745]: E1010 13:19:08.744821 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.839951 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.839985 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.839994 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.840007 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.840016 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:08Z","lastTransitionTime":"2025-10-10T13:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.942579 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.942642 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.942659 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.942686 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:08 crc kubenswrapper[4745]: I1010 13:19:08.942705 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:08Z","lastTransitionTime":"2025-10-10T13:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.045557 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.045627 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.045639 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.045658 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.045671 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:09Z","lastTransitionTime":"2025-10-10T13:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.148979 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.149029 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.149043 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.149063 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.149077 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:09Z","lastTransitionTime":"2025-10-10T13:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.252560 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.252622 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.252640 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.252664 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.252681 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:09Z","lastTransitionTime":"2025-10-10T13:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.355762 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.355817 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.355833 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.355854 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.355871 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:09Z","lastTransitionTime":"2025-10-10T13:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.458940 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.458992 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.459008 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.459030 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.459048 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:09Z","lastTransitionTime":"2025-10-10T13:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.562048 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.562102 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.562119 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.562142 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.562159 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:09Z","lastTransitionTime":"2025-10-10T13:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.665130 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.665197 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.665215 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.665239 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.665256 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:09Z","lastTransitionTime":"2025-10-10T13:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.745038 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.745124 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:09 crc kubenswrapper[4745]: E1010 13:19:09.745234 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:09 crc kubenswrapper[4745]: E1010 13:19:09.745354 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.768603 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.768658 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.768676 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.768702 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.768722 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:09Z","lastTransitionTime":"2025-10-10T13:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.872291 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.872351 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.872371 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.872397 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.872415 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:09Z","lastTransitionTime":"2025-10-10T13:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.975466 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.975537 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.975571 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.975603 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:09 crc kubenswrapper[4745]: I1010 13:19:09.975627 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:09Z","lastTransitionTime":"2025-10-10T13:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.079409 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.079485 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.079508 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.079539 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.079562 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:10Z","lastTransitionTime":"2025-10-10T13:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.182264 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.182337 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.182360 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.182384 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.182402 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:10Z","lastTransitionTime":"2025-10-10T13:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.285127 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.285184 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.285200 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.285222 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.285238 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:10Z","lastTransitionTime":"2025-10-10T13:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.388027 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.388067 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.388080 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.388098 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.388111 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:10Z","lastTransitionTime":"2025-10-10T13:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.490473 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.490509 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.490520 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.490535 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.490547 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:10Z","lastTransitionTime":"2025-10-10T13:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.592945 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.593005 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.593022 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.593043 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.593058 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:10Z","lastTransitionTime":"2025-10-10T13:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.695798 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.695854 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.695874 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.695897 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.695917 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:10Z","lastTransitionTime":"2025-10-10T13:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.744862 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.744902 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:10 crc kubenswrapper[4745]: E1010 13:19:10.745098 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:10 crc kubenswrapper[4745]: E1010 13:19:10.745213 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.798869 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.799231 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.799427 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.799580 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.799758 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:10Z","lastTransitionTime":"2025-10-10T13:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.902354 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.903331 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.903578 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.903775 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:10 crc kubenswrapper[4745]: I1010 13:19:10.904096 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:10Z","lastTransitionTime":"2025-10-10T13:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.007449 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.007509 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.007529 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.007556 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.007574 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:11Z","lastTransitionTime":"2025-10-10T13:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.110052 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.110090 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.110101 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.110157 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.110171 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:11Z","lastTransitionTime":"2025-10-10T13:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.213312 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.213364 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.213376 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.213392 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.213403 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:11Z","lastTransitionTime":"2025-10-10T13:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.316795 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.316835 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.316847 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.316864 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.316875 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:11Z","lastTransitionTime":"2025-10-10T13:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.419797 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.419836 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.419844 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.419860 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.419869 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:11Z","lastTransitionTime":"2025-10-10T13:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.522792 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.522829 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.522885 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.522904 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.522917 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:11Z","lastTransitionTime":"2025-10-10T13:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.626669 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.627178 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.627328 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.627483 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.627614 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:11Z","lastTransitionTime":"2025-10-10T13:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.731064 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.731347 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.731509 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.731645 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.731808 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:11Z","lastTransitionTime":"2025-10-10T13:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.744595 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:11 crc kubenswrapper[4745]: E1010 13:19:11.744867 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.744895 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:11 crc kubenswrapper[4745]: E1010 13:19:11.745123 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.835007 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.835284 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.835387 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.835474 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.835555 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:11Z","lastTransitionTime":"2025-10-10T13:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.937522 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.937545 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.937552 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.937565 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:11 crc kubenswrapper[4745]: I1010 13:19:11.937574 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:11Z","lastTransitionTime":"2025-10-10T13:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.040019 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.040099 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.040125 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.040156 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.040177 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:12Z","lastTransitionTime":"2025-10-10T13:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.143370 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.143420 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.143432 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.143449 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.143461 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:12Z","lastTransitionTime":"2025-10-10T13:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.245828 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.245875 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.245887 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.245905 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.245917 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:12Z","lastTransitionTime":"2025-10-10T13:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.348399 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.348440 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.348449 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.348463 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.348475 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:12Z","lastTransitionTime":"2025-10-10T13:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.450793 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.450833 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.450841 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.450856 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.450869 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:12Z","lastTransitionTime":"2025-10-10T13:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.553917 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.553963 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.553976 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.553992 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.554004 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:12Z","lastTransitionTime":"2025-10-10T13:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.656057 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.656086 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.656096 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.656107 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.656118 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:12Z","lastTransitionTime":"2025-10-10T13:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.744753 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:12 crc kubenswrapper[4745]: E1010 13:19:12.744923 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.745175 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:12 crc kubenswrapper[4745]: E1010 13:19:12.745259 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.758616 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.758646 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.758657 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.758671 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.758682 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:12Z","lastTransitionTime":"2025-10-10T13:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.861817 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.861876 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.861887 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.861902 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.861913 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:12Z","lastTransitionTime":"2025-10-10T13:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.964476 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.964525 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.964535 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.964550 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:12 crc kubenswrapper[4745]: I1010 13:19:12.964560 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:12Z","lastTransitionTime":"2025-10-10T13:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.066548 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.066597 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.066609 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.066626 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.066640 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:13Z","lastTransitionTime":"2025-10-10T13:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.169681 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.169756 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.169768 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.169790 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.169802 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:13Z","lastTransitionTime":"2025-10-10T13:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.272188 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.272243 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.272251 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.272266 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.272275 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:13Z","lastTransitionTime":"2025-10-10T13:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.374472 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.374505 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.374513 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.374527 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.374535 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:13Z","lastTransitionTime":"2025-10-10T13:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.476546 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.476594 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.476602 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.476615 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.476624 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:13Z","lastTransitionTime":"2025-10-10T13:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.579179 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.579212 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.579220 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.579234 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.579242 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:13Z","lastTransitionTime":"2025-10-10T13:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.681704 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.681777 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.681789 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.681812 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.681823 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:13Z","lastTransitionTime":"2025-10-10T13:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.744092 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.744170 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:13 crc kubenswrapper[4745]: E1010 13:19:13.744290 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:13 crc kubenswrapper[4745]: E1010 13:19:13.744559 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.783578 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.783633 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.783646 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.783664 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.783675 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:13Z","lastTransitionTime":"2025-10-10T13:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.886604 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.886659 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.886676 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.886706 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.886768 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:13Z","lastTransitionTime":"2025-10-10T13:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.988618 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.988653 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.988664 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.988679 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:13 crc kubenswrapper[4745]: I1010 13:19:13.988689 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:13Z","lastTransitionTime":"2025-10-10T13:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.007300 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs\") pod \"network-metrics-daemon-xgh8p\" (UID: \"82c4e63a-009f-4720-87b5-640a8ec65117\") " pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:14 crc kubenswrapper[4745]: E1010 13:19:14.007466 4745 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 13:19:14 crc kubenswrapper[4745]: E1010 13:19:14.007527 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs podName:82c4e63a-009f-4720-87b5-640a8ec65117 nodeName:}" failed. No retries permitted until 2025-10-10 13:19:46.007511567 +0000 UTC m=+99.905168330 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs") pod "network-metrics-daemon-xgh8p" (UID: "82c4e63a-009f-4720-87b5-640a8ec65117") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.091566 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.091619 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.091631 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.091648 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.091662 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:14Z","lastTransitionTime":"2025-10-10T13:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.193717 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.193777 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.193789 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.193807 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.193818 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:14Z","lastTransitionTime":"2025-10-10T13:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.222810 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.222839 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.222849 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.222864 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.222876 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:14Z","lastTransitionTime":"2025-10-10T13:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:14 crc kubenswrapper[4745]: E1010 13:19:14.239220 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:14Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.242859 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.242886 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.242895 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.242907 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.242918 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:14Z","lastTransitionTime":"2025-10-10T13:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:14 crc kubenswrapper[4745]: E1010 13:19:14.257984 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:14Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.261701 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.261784 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.261802 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.261825 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.261844 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:14Z","lastTransitionTime":"2025-10-10T13:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:14 crc kubenswrapper[4745]: E1010 13:19:14.275456 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:14Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.279400 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.279457 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.279477 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.279500 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.279518 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:14Z","lastTransitionTime":"2025-10-10T13:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:14 crc kubenswrapper[4745]: E1010 13:19:14.293237 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:14Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.297295 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.297346 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.297362 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.297384 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.297401 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:14Z","lastTransitionTime":"2025-10-10T13:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:14 crc kubenswrapper[4745]: E1010 13:19:14.311557 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:14Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:14 crc kubenswrapper[4745]: E1010 13:19:14.311789 4745 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.314003 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.314043 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.314054 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.314070 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.314081 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:14Z","lastTransitionTime":"2025-10-10T13:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.416303 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.416342 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.416354 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.416369 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.416381 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:14Z","lastTransitionTime":"2025-10-10T13:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.518845 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.518891 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.518903 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.518920 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.518933 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:14Z","lastTransitionTime":"2025-10-10T13:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.621412 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.621472 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.621495 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.621559 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.621582 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:14Z","lastTransitionTime":"2025-10-10T13:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.723954 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.724038 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.724055 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.724079 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.724100 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:14Z","lastTransitionTime":"2025-10-10T13:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.744968 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.745072 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:14 crc kubenswrapper[4745]: E1010 13:19:14.745603 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:14 crc kubenswrapper[4745]: E1010 13:19:14.746122 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.826959 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.826991 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.827001 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.827018 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.827029 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:14Z","lastTransitionTime":"2025-10-10T13:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.929419 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.929459 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.929472 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.929489 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:14 crc kubenswrapper[4745]: I1010 13:19:14.929500 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:14Z","lastTransitionTime":"2025-10-10T13:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.032486 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.032565 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.032587 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.032617 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.032639 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:15Z","lastTransitionTime":"2025-10-10T13:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.135896 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.135946 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.135959 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.135978 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.135994 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:15Z","lastTransitionTime":"2025-10-10T13:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.238429 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.238491 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.238508 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.238532 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.238551 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:15Z","lastTransitionTime":"2025-10-10T13:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.341398 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.341463 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.341481 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.341507 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.341527 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:15Z","lastTransitionTime":"2025-10-10T13:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.445151 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.445204 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.445221 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.445248 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.445266 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:15Z","lastTransitionTime":"2025-10-10T13:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.547709 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.547796 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.547814 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.547839 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.547858 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:15Z","lastTransitionTime":"2025-10-10T13:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.651522 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.651555 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.651566 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.651581 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.651592 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:15Z","lastTransitionTime":"2025-10-10T13:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.744819 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.744830 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:15 crc kubenswrapper[4745]: E1010 13:19:15.744964 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:15 crc kubenswrapper[4745]: E1010 13:19:15.745045 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.753671 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.753718 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.753744 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.753763 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.753774 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:15Z","lastTransitionTime":"2025-10-10T13:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.857091 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.857137 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.857148 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.857164 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.857176 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:15Z","lastTransitionTime":"2025-10-10T13:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.960057 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.960112 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.960125 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.960143 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:15 crc kubenswrapper[4745]: I1010 13:19:15.960156 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:15Z","lastTransitionTime":"2025-10-10T13:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.062503 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.062560 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.062578 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.062605 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.062625 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:16Z","lastTransitionTime":"2025-10-10T13:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.164483 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.164510 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.164520 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.164536 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.164547 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:16Z","lastTransitionTime":"2025-10-10T13:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.194268 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-27d2n_1d62a30b-50a8-46a2-82c6-5d7e2f24e4de/kube-multus/0.log" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.194318 4745 generic.go:334] "Generic (PLEG): container finished" podID="1d62a30b-50a8-46a2-82c6-5d7e2f24e4de" containerID="9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb" exitCode=1 Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.194347 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-27d2n" event={"ID":"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de","Type":"ContainerDied","Data":"9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb"} Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.194752 4745 scope.go:117] "RemoveContainer" containerID="9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.208934 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82c4e63a-009f-4720-87b5-640a8ec65117\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xgh8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.233230 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923c385c-8102-4892-a4a3-67b6408545ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://193833f7e032a32aae576702c6965f7ea2869114b86d1a0349e7ad787869c74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab4465f351419e32376cc0b2062225c6faac1e27dce51bcf3c1ca4bb590fc4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f6659c3fefc2b4892b34a407a26418aa86c780a7e03c9120794f4396c08913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.253017 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.268179 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.268216 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.268228 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.268265 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.268279 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:16Z","lastTransitionTime":"2025-10-10T13:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.272289 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:19:15Z\\\",\\\"message\\\":\\\"2025-10-10T13:18:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e6a3b6a9-97e0-4aad-9ab9-eb4f852eb8d1\\\\n2025-10-10T13:18:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e6a3b6a9-97e0-4aad-9ab9-eb4f852eb8d1 to /host/opt/cni/bin/\\\\n2025-10-10T13:18:30Z [verbose] multus-daemon started\\\\n2025-10-10T13:18:30Z [verbose] Readiness Indicator file check\\\\n2025-10-10T13:19:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.291007 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.314943 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.331396 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.351996 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:51Z\\\",\\\"message\\\":\\\"6 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 13:18:51.354210 6376 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 13:18:51.354660 6376 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354695 6376 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354832 6376 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355149 6376 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355386 6376 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1010 13:18:51.355631 6376 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.366100 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.370230 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.370293 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.370304 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.370317 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.370326 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:16Z","lastTransitionTime":"2025-10-10T13:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.389848 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.405409 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.417843 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.428423 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.437026 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.443910 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.461429 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.471836 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.472718 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.472757 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.472765 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.472779 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.472788 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:16Z","lastTransitionTime":"2025-10-10T13:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.492974 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.574647 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.574884 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.574892 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.574907 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.574917 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:16Z","lastTransitionTime":"2025-10-10T13:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.677935 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.677971 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.677982 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.677996 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.678005 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:16Z","lastTransitionTime":"2025-10-10T13:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.744922 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.745001 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:16 crc kubenswrapper[4745]: E1010 13:19:16.745103 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:16 crc kubenswrapper[4745]: E1010 13:19:16.745168 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.761704 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.779545 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.780780 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.780813 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.780825 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.780844 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.780856 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:16Z","lastTransitionTime":"2025-10-10T13:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.792445 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.806977 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.821073 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.836250 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.849220 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923c385c-8102-4892-a4a3-67b6408545ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://193833f7e032a32aae576702c6965f7ea2869114b86d1a0349e7ad787869c74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab4465f351419e32376cc0b2062225c6faac1e27dce51bcf3c1ca4bb590fc4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f6659c3fefc2b4892b34a407a26418aa86c780a7e03c9120794f4396c08913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.861953 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.879378 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:19:15Z\\\",\\\"message\\\":\\\"2025-10-10T13:18:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e6a3b6a9-97e0-4aad-9ab9-eb4f852eb8d1\\\\n2025-10-10T13:18:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e6a3b6a9-97e0-4aad-9ab9-eb4f852eb8d1 to /host/opt/cni/bin/\\\\n2025-10-10T13:18:30Z [verbose] multus-daemon started\\\\n2025-10-10T13:18:30Z [verbose] Readiness Indicator file check\\\\n2025-10-10T13:19:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.883222 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.883267 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.883285 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.883309 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.883326 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:16Z","lastTransitionTime":"2025-10-10T13:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.892629 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.904097 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82c4e63a-009f-4720-87b5-640a8ec65117\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xgh8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.916222 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.948792 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:51Z\\\",\\\"message\\\":\\\"6 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 13:18:51.354210 6376 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 13:18:51.354660 6376 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354695 6376 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354832 6376 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355149 6376 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355386 6376 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1010 13:18:51.355631 6376 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.982663 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:16Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.985105 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.985136 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.985147 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.985166 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:16 crc kubenswrapper[4745]: I1010 13:19:16.985178 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:16Z","lastTransitionTime":"2025-10-10T13:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.007344 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.021538 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.033040 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.045191 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.088365 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.088407 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.088417 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.088435 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.088445 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:17Z","lastTransitionTime":"2025-10-10T13:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.190749 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.190787 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.190796 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.190812 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.190822 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:17Z","lastTransitionTime":"2025-10-10T13:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.198270 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-27d2n_1d62a30b-50a8-46a2-82c6-5d7e2f24e4de/kube-multus/0.log" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.198403 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-27d2n" event={"ID":"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de","Type":"ContainerStarted","Data":"2f78611b429d8411902427cc4f7dda27fb9f1ea97db3b312737d4bbc38d350a8"} Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.212150 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.228258 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.253038 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.270715 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.287292 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f78611b429d8411902427cc4f7dda27fb9f1ea97db3b312737d4bbc38d350a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:19:15Z\\\",\\\"message\\\":\\\"2025-10-10T13:18:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e6a3b6a9-97e0-4aad-9ab9-eb4f852eb8d1\\\\n2025-10-10T13:18:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e6a3b6a9-97e0-4aad-9ab9-eb4f852eb8d1 to /host/opt/cni/bin/\\\\n2025-10-10T13:18:30Z [verbose] multus-daemon started\\\\n2025-10-10T13:18:30Z [verbose] Readiness Indicator file check\\\\n2025-10-10T13:19:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:19:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.293682 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.293746 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.293762 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.293780 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.293795 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:17Z","lastTransitionTime":"2025-10-10T13:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.301550 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.312873 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82c4e63a-009f-4720-87b5-640a8ec65117\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xgh8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.326020 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923c385c-8102-4892-a4a3-67b6408545ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://193833f7e032a32aae576702c6965f7ea2869114b86d1a0349e7ad787869c74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab4465f351419e32376cc0b2062225c6faac1e27dce51bcf3c1ca4bb590fc4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f6659c3fefc2b4892b34a407a26418aa86c780a7e03c9120794f4396c08913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.344024 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.361668 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.376046 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.389694 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.396637 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.396691 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.396710 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.396763 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.396783 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:17Z","lastTransitionTime":"2025-10-10T13:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.398650 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.419275 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:51Z\\\",\\\"message\\\":\\\"6 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 13:18:51.354210 6376 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 13:18:51.354660 6376 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354695 6376 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354832 6376 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355149 6376 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355386 6376 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1010 13:18:51.355631 6376 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.434655 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.447556 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.461510 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.475140 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:17Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.499252 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.499323 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.499342 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.499367 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.499386 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:17Z","lastTransitionTime":"2025-10-10T13:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.602368 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.602406 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.602417 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.602432 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.602442 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:17Z","lastTransitionTime":"2025-10-10T13:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.705100 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.705129 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.705138 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.705150 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.705160 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:17Z","lastTransitionTime":"2025-10-10T13:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.744458 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:17 crc kubenswrapper[4745]: E1010 13:19:17.744630 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.744689 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:17 crc kubenswrapper[4745]: E1010 13:19:17.745091 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.808235 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.808295 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.808313 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.808338 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.808357 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:17Z","lastTransitionTime":"2025-10-10T13:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.910856 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.910914 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.910935 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.910961 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:17 crc kubenswrapper[4745]: I1010 13:19:17.910979 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:17Z","lastTransitionTime":"2025-10-10T13:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.013615 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.013655 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.013666 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.013687 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.013699 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:18Z","lastTransitionTime":"2025-10-10T13:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.116464 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.116501 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.116510 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.116541 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.116552 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:18Z","lastTransitionTime":"2025-10-10T13:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.219045 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.219076 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.219085 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.219097 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.219106 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:18Z","lastTransitionTime":"2025-10-10T13:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.322418 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.322468 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.322485 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.322510 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.322538 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:18Z","lastTransitionTime":"2025-10-10T13:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.425615 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.425661 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.425673 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.425691 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.425704 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:18Z","lastTransitionTime":"2025-10-10T13:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.528100 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.528155 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.528168 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.528184 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.528196 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:18Z","lastTransitionTime":"2025-10-10T13:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.629828 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.629892 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.629913 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.629945 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.629968 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:18Z","lastTransitionTime":"2025-10-10T13:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.732293 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.732334 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.732345 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.732362 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.732373 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:18Z","lastTransitionTime":"2025-10-10T13:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.744575 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:18 crc kubenswrapper[4745]: E1010 13:19:18.744663 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.744703 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:18 crc kubenswrapper[4745]: E1010 13:19:18.745161 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.745515 4745 scope.go:117] "RemoveContainer" containerID="7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.835214 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.835259 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.835274 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.835299 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.835315 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:18Z","lastTransitionTime":"2025-10-10T13:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.937176 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.937248 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.937266 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.937773 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:18 crc kubenswrapper[4745]: I1010 13:19:18.937832 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:18Z","lastTransitionTime":"2025-10-10T13:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.044584 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.044639 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.044653 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.044670 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.044686 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:19Z","lastTransitionTime":"2025-10-10T13:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.147865 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.147944 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.147961 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.148273 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.148308 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:19Z","lastTransitionTime":"2025-10-10T13:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.206243 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovnkube-controller/2.log" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.209252 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerStarted","Data":"b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8"} Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.210205 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.220332 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.238685 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:51Z\\\",\\\"message\\\":\\\"6 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 13:18:51.354210 6376 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 13:18:51.354660 6376 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354695 6376 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354832 6376 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355149 6376 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355386 6376 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1010 13:18:51.355631 6376 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:19:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.250418 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.250458 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.250469 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.250487 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.250499 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:19Z","lastTransitionTime":"2025-10-10T13:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.251447 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.272392 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.287273 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.299423 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.313855 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.332024 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.352787 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.352840 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.352848 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.352883 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.352892 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:19Z","lastTransitionTime":"2025-10-10T13:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.357062 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.372396 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.389323 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.412501 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.428327 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.439350 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923c385c-8102-4892-a4a3-67b6408545ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://193833f7e032a32aae576702c6965f7ea2869114b86d1a0349e7ad787869c74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab4465f351419e32376cc0b2062225c6faac1e27dce51bcf3c1ca4bb590fc4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f6659c3fefc2b4892b34a407a26418aa86c780a7e03c9120794f4396c08913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.449214 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.454530 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.454569 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.454603 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.454621 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.454633 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:19Z","lastTransitionTime":"2025-10-10T13:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.461989 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f78611b429d8411902427cc4f7dda27fb9f1ea97db3b312737d4bbc38d350a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:19:15Z\\\",\\\"message\\\":\\\"2025-10-10T13:18:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e6a3b6a9-97e0-4aad-9ab9-eb4f852eb8d1\\\\n2025-10-10T13:18:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e6a3b6a9-97e0-4aad-9ab9-eb4f852eb8d1 to /host/opt/cni/bin/\\\\n2025-10-10T13:18:30Z [verbose] multus-daemon started\\\\n2025-10-10T13:18:30Z [verbose] Readiness Indicator file check\\\\n2025-10-10T13:19:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:19:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.473064 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.484322 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82c4e63a-009f-4720-87b5-640a8ec65117\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xgh8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.557042 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.557081 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.557089 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.557103 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.557133 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:19Z","lastTransitionTime":"2025-10-10T13:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.660035 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.660079 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.660091 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.660105 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.660114 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:19Z","lastTransitionTime":"2025-10-10T13:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.744960 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.744989 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:19 crc kubenswrapper[4745]: E1010 13:19:19.745105 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:19 crc kubenswrapper[4745]: E1010 13:19:19.745251 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.763225 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.763265 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.763279 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.763298 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.763314 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:19Z","lastTransitionTime":"2025-10-10T13:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.865327 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.865366 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.865377 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.865390 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.865400 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:19Z","lastTransitionTime":"2025-10-10T13:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.967769 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.968021 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.968037 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.968059 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:19 crc kubenswrapper[4745]: I1010 13:19:19.968077 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:19Z","lastTransitionTime":"2025-10-10T13:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.071387 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.071493 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.071514 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.071540 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.071557 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:20Z","lastTransitionTime":"2025-10-10T13:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.173628 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.173707 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.173765 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.173790 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.173807 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:20Z","lastTransitionTime":"2025-10-10T13:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.214623 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovnkube-controller/3.log" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.215558 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovnkube-controller/2.log" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.218068 4745 generic.go:334] "Generic (PLEG): container finished" podID="687bad4e-baff-4e09-bb33-def12c499092" containerID="b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8" exitCode=1 Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.218118 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerDied","Data":"b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8"} Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.218159 4745 scope.go:117] "RemoveContainer" containerID="7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.218814 4745 scope.go:117] "RemoveContainer" containerID="b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8" Oct 10 13:19:20 crc kubenswrapper[4745]: E1010 13:19:20.218974 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podUID="687bad4e-baff-4e09-bb33-def12c499092" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.229071 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.241694 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.255095 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.267869 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.279225 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.279262 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.279271 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.279286 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.279296 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:20Z","lastTransitionTime":"2025-10-10T13:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.283665 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.304487 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.320002 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.334201 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82c4e63a-009f-4720-87b5-640a8ec65117\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xgh8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.351609 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923c385c-8102-4892-a4a3-67b6408545ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://193833f7e032a32aae576702c6965f7ea2869114b86d1a0349e7ad787869c74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab4465f351419e32376cc0b2062225c6faac1e27dce51bcf3c1ca4bb590fc4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f6659c3fefc2b4892b34a407a26418aa86c780a7e03c9120794f4396c08913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.366455 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.382392 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.382426 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.382438 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.382452 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.382462 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:20Z","lastTransitionTime":"2025-10-10T13:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.384208 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f78611b429d8411902427cc4f7dda27fb9f1ea97db3b312737d4bbc38d350a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:19:15Z\\\",\\\"message\\\":\\\"2025-10-10T13:18:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e6a3b6a9-97e0-4aad-9ab9-eb4f852eb8d1\\\\n2025-10-10T13:18:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e6a3b6a9-97e0-4aad-9ab9-eb4f852eb8d1 to /host/opt/cni/bin/\\\\n2025-10-10T13:18:30Z [verbose] multus-daemon started\\\\n2025-10-10T13:18:30Z [verbose] Readiness Indicator file check\\\\n2025-10-10T13:19:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:19:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.400424 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.415981 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.430586 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.452988 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7487a7ce846297f7c28666a016dcf1d3c9dc3b831aff0daa64ed497a48e99d3e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:18:51Z\\\",\\\"message\\\":\\\"6 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 13:18:51.354210 6376 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 13:18:51.354660 6376 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354695 6376 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.354832 6376 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355149 6376 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 13:18:51.355386 6376 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1010 13:18:51.355631 6376 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:19:19Z\\\",\\\"message\\\":\\\" not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z]\\\\nI1010 13:19:19.610281 6728 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1010 13:19:19.610285 6728 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 13:19:19.610282 6728 obj_retry.go:285] Attempting retry of *v1.Pod openshift-multus/network-metrics-daemon-xgh8p before timer (time: 2025-10-10 13:19:20.777105376 +0000 UTC m=+1.753565904): skip\\\\nI1010 13:19:19.610292 6728 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1010 13:19:19.610295 6728 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1010 13:19:19.610297 6728 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 13:19:19.610300 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:19:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.464953 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.484332 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.484371 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.484382 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.484398 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.484410 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:20Z","lastTransitionTime":"2025-10-10T13:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.484799 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.499369 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:20Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.588146 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.588241 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.588262 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.588297 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.588320 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:20Z","lastTransitionTime":"2025-10-10T13:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.691258 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.691322 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.691338 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.691362 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.691381 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:20Z","lastTransitionTime":"2025-10-10T13:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.744464 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.744466 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:20 crc kubenswrapper[4745]: E1010 13:19:20.744683 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:20 crc kubenswrapper[4745]: E1010 13:19:20.744892 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.794593 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.794653 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.794670 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.794693 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.794711 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:20Z","lastTransitionTime":"2025-10-10T13:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.897845 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.897917 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.897941 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.897971 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:20 crc kubenswrapper[4745]: I1010 13:19:20.897998 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:20Z","lastTransitionTime":"2025-10-10T13:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.001092 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.001157 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.001175 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.001199 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.001218 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:21Z","lastTransitionTime":"2025-10-10T13:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.105186 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.105284 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.105309 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.105338 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.105359 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:21Z","lastTransitionTime":"2025-10-10T13:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.208933 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.209037 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.209063 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.209139 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.209159 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:21Z","lastTransitionTime":"2025-10-10T13:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.224484 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovnkube-controller/3.log" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.229711 4745 scope.go:117] "RemoveContainer" containerID="b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8" Oct 10 13:19:21 crc kubenswrapper[4745]: E1010 13:19:21.230001 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podUID="687bad4e-baff-4e09-bb33-def12c499092" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.251388 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.270137 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.291442 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.306982 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923c385c-8102-4892-a4a3-67b6408545ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://193833f7e032a32aae576702c6965f7ea2869114b86d1a0349e7ad787869c74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab4465f351419e32376cc0b2062225c6faac1e27dce51bcf3c1ca4bb590fc4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f6659c3fefc2b4892b34a407a26418aa86c780a7e03c9120794f4396c08913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.313038 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.313083 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.313106 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.313127 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.313142 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:21Z","lastTransitionTime":"2025-10-10T13:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.322642 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.336436 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f78611b429d8411902427cc4f7dda27fb9f1ea97db3b312737d4bbc38d350a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:19:15Z\\\",\\\"message\\\":\\\"2025-10-10T13:18:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e6a3b6a9-97e0-4aad-9ab9-eb4f852eb8d1\\\\n2025-10-10T13:18:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e6a3b6a9-97e0-4aad-9ab9-eb4f852eb8d1 to /host/opt/cni/bin/\\\\n2025-10-10T13:18:30Z [verbose] multus-daemon started\\\\n2025-10-10T13:18:30Z [verbose] Readiness Indicator file check\\\\n2025-10-10T13:19:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:19:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.350621 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.362186 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82c4e63a-009f-4720-87b5-640a8ec65117\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xgh8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.377585 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.409177 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.415256 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.415287 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.415299 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.415316 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.415329 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:21Z","lastTransitionTime":"2025-10-10T13:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.428684 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.442978 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.460376 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.473552 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.496401 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:19:19Z\\\",\\\"message\\\":\\\" not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z]\\\\nI1010 13:19:19.610281 6728 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1010 13:19:19.610285 6728 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 13:19:19.610282 6728 obj_retry.go:285] Attempting retry of *v1.Pod openshift-multus/network-metrics-daemon-xgh8p before timer (time: 2025-10-10 13:19:20.777105376 +0000 UTC m=+1.753565904): skip\\\\nI1010 13:19:19.610292 6728 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1010 13:19:19.610295 6728 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1010 13:19:19.610297 6728 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 13:19:19.610300 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:19:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.508896 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.517382 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.517426 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.517437 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.517453 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.517464 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:21Z","lastTransitionTime":"2025-10-10T13:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.519667 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.528827 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:21Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.620564 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.620616 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.620630 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.620651 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.620664 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:21Z","lastTransitionTime":"2025-10-10T13:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.723263 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.723343 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.723368 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.723400 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.723423 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:21Z","lastTransitionTime":"2025-10-10T13:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.744265 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.744284 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:21 crc kubenswrapper[4745]: E1010 13:19:21.744497 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:21 crc kubenswrapper[4745]: E1010 13:19:21.744649 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.759339 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.825823 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.825898 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.825916 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.825954 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.825975 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:21Z","lastTransitionTime":"2025-10-10T13:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.928509 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.928563 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.928580 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.928603 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:21 crc kubenswrapper[4745]: I1010 13:19:21.928622 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:21Z","lastTransitionTime":"2025-10-10T13:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.031928 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.031992 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.032010 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.032033 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.032050 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:22Z","lastTransitionTime":"2025-10-10T13:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.134865 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.134902 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.134914 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.134932 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.134944 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:22Z","lastTransitionTime":"2025-10-10T13:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.238194 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.238273 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.238291 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.238317 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.238335 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:22Z","lastTransitionTime":"2025-10-10T13:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.341925 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.341979 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.341996 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.342021 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.342037 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:22Z","lastTransitionTime":"2025-10-10T13:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.444990 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.445062 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.445079 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.445105 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.445124 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:22Z","lastTransitionTime":"2025-10-10T13:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.548719 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.548829 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.548850 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.548874 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.548905 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:22Z","lastTransitionTime":"2025-10-10T13:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.652702 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.652792 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.652810 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.652834 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.652850 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:22Z","lastTransitionTime":"2025-10-10T13:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.744552 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:22 crc kubenswrapper[4745]: E1010 13:19:22.744712 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.745043 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:22 crc kubenswrapper[4745]: E1010 13:19:22.745197 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.754943 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.754987 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.755001 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.755018 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.755034 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:22Z","lastTransitionTime":"2025-10-10T13:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.858010 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.858079 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.858107 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.858138 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.858157 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:22Z","lastTransitionTime":"2025-10-10T13:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.960788 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.960862 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.960890 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.960934 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:22 crc kubenswrapper[4745]: I1010 13:19:22.960958 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:22Z","lastTransitionTime":"2025-10-10T13:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.063388 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.063435 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.063445 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.063460 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.063471 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:23Z","lastTransitionTime":"2025-10-10T13:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.165766 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.165819 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.165834 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.165852 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.165864 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:23Z","lastTransitionTime":"2025-10-10T13:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.268588 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.268663 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.268685 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.268716 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.268773 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:23Z","lastTransitionTime":"2025-10-10T13:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.370800 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.370869 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.370892 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.370919 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.370944 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:23Z","lastTransitionTime":"2025-10-10T13:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.477975 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.478549 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.478585 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.478610 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.478627 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:23Z","lastTransitionTime":"2025-10-10T13:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.581612 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.581675 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.581688 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.581703 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.581715 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:23Z","lastTransitionTime":"2025-10-10T13:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.685417 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.685503 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.685522 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.685546 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.685593 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:23Z","lastTransitionTime":"2025-10-10T13:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.744917 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.745175 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:23 crc kubenswrapper[4745]: E1010 13:19:23.745353 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:23 crc kubenswrapper[4745]: E1010 13:19:23.745514 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.787892 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.787955 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.787966 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.787983 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.788026 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:23Z","lastTransitionTime":"2025-10-10T13:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.890442 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.890496 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.890515 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.890538 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.890556 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:23Z","lastTransitionTime":"2025-10-10T13:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.993382 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.993431 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.993449 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.993473 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:23 crc kubenswrapper[4745]: I1010 13:19:23.993496 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:23Z","lastTransitionTime":"2025-10-10T13:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.096609 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.096664 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.096680 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.096707 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.096768 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:24Z","lastTransitionTime":"2025-10-10T13:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.199774 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.199818 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.199834 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.199856 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.199872 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:24Z","lastTransitionTime":"2025-10-10T13:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.302257 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.302291 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.302303 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.302318 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.302329 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:24Z","lastTransitionTime":"2025-10-10T13:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.404912 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.404977 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.404996 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.405021 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.405039 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:24Z","lastTransitionTime":"2025-10-10T13:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.508149 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.508192 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.508204 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.508257 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.508271 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:24Z","lastTransitionTime":"2025-10-10T13:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.550119 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.550165 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.550177 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.550198 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.550212 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:24Z","lastTransitionTime":"2025-10-10T13:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:24 crc kubenswrapper[4745]: E1010 13:19:24.566383 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:24Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.570797 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.570838 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.570851 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.570868 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.570879 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:24Z","lastTransitionTime":"2025-10-10T13:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:24 crc kubenswrapper[4745]: E1010 13:19:24.588229 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:24Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.592504 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.592532 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.592542 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.592558 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.592571 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:24Z","lastTransitionTime":"2025-10-10T13:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:24 crc kubenswrapper[4745]: E1010 13:19:24.610044 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:24Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.614834 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.614871 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.614879 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.614893 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.614902 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:24Z","lastTransitionTime":"2025-10-10T13:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:24 crc kubenswrapper[4745]: E1010 13:19:24.626365 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:24Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.630343 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.630378 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.630388 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.630401 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.630410 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:24Z","lastTransitionTime":"2025-10-10T13:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:24 crc kubenswrapper[4745]: E1010 13:19:24.648604 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:24Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:24 crc kubenswrapper[4745]: E1010 13:19:24.648818 4745 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.650273 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.650330 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.650351 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.650374 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.650393 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:24Z","lastTransitionTime":"2025-10-10T13:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.744398 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.744662 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:24 crc kubenswrapper[4745]: E1010 13:19:24.744617 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:24 crc kubenswrapper[4745]: E1010 13:19:24.744854 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.753037 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.753068 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.753078 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.753092 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.753103 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:24Z","lastTransitionTime":"2025-10-10T13:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.855681 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.855716 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.855724 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.855754 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.855763 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:24Z","lastTransitionTime":"2025-10-10T13:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.958048 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.958113 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.958135 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.958168 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:24 crc kubenswrapper[4745]: I1010 13:19:24.958190 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:24Z","lastTransitionTime":"2025-10-10T13:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.061651 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.061763 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.061792 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.061823 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.061849 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:25Z","lastTransitionTime":"2025-10-10T13:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.164724 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.164789 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.164800 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.164816 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.164828 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:25Z","lastTransitionTime":"2025-10-10T13:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.268528 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.268578 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.268587 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.268601 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.268613 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:25Z","lastTransitionTime":"2025-10-10T13:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.371676 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.371719 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.371740 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.371798 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.371807 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:25Z","lastTransitionTime":"2025-10-10T13:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.474004 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.474038 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.474046 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.474059 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.474067 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:25Z","lastTransitionTime":"2025-10-10T13:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.576132 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.576172 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.576183 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.576200 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.576212 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:25Z","lastTransitionTime":"2025-10-10T13:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.678918 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.678959 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.678973 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.678990 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.679003 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:25Z","lastTransitionTime":"2025-10-10T13:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.744259 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.744300 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:25 crc kubenswrapper[4745]: E1010 13:19:25.744657 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:25 crc kubenswrapper[4745]: E1010 13:19:25.744557 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.781101 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.781140 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.781154 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.781173 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.781187 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:25Z","lastTransitionTime":"2025-10-10T13:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.883453 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.883488 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.883497 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.883514 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.883524 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:25Z","lastTransitionTime":"2025-10-10T13:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.986116 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.986170 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.986184 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.986203 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:25 crc kubenswrapper[4745]: I1010 13:19:25.986220 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:25Z","lastTransitionTime":"2025-10-10T13:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.089132 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.089189 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.089205 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.089225 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.089239 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:26Z","lastTransitionTime":"2025-10-10T13:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.191147 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.191447 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.191542 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.191635 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.191758 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:26Z","lastTransitionTime":"2025-10-10T13:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.294178 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.294471 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.294596 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.294685 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.294789 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:26Z","lastTransitionTime":"2025-10-10T13:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.397594 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.397672 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.397690 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.397713 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.397760 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:26Z","lastTransitionTime":"2025-10-10T13:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.500359 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.500643 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.500654 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.500672 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.500685 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:26Z","lastTransitionTime":"2025-10-10T13:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.602696 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.602784 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.602799 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.602818 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.602830 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:26Z","lastTransitionTime":"2025-10-10T13:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.704797 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.704833 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.704851 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.704872 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.704883 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:26Z","lastTransitionTime":"2025-10-10T13:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.745022 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:26 crc kubenswrapper[4745]: E1010 13:19:26.745148 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.745022 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:26 crc kubenswrapper[4745]: E1010 13:19:26.745407 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.766188 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f78611b429d8411902427cc4f7dda27fb9f1ea97db3b312737d4bbc38d350a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:19:15Z\\\",\\\"message\\\":\\\"2025-10-10T13:18:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e6a3b6a9-97e0-4aad-9ab9-eb4f852eb8d1\\\\n2025-10-10T13:18:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e6a3b6a9-97e0-4aad-9ab9-eb4f852eb8d1 to /host/opt/cni/bin/\\\\n2025-10-10T13:18:30Z [verbose] multus-daemon started\\\\n2025-10-10T13:18:30Z [verbose] Readiness Indicator file check\\\\n2025-10-10T13:19:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:19:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:26Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.783856 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:26Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.800894 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82c4e63a-009f-4720-87b5-640a8ec65117\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xgh8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:26Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.807122 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.807153 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.807161 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.807174 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.807183 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:26Z","lastTransitionTime":"2025-10-10T13:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.814860 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923c385c-8102-4892-a4a3-67b6408545ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://193833f7e032a32aae576702c6965f7ea2869114b86d1a0349e7ad787869c74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab4465f351419e32376cc0b2062225c6faac1e27dce51bcf3c1ca4bb590fc4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f6659c3fefc2b4892b34a407a26418aa86c780a7e03c9120794f4396c08913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:26Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.825319 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:26Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.844287 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:26Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.856448 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:26Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.877058 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:26Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.891203 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:26Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.909799 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.910069 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.910176 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.910362 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.910877 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:26Z","lastTransitionTime":"2025-10-10T13:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.912075 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:19:19Z\\\",\\\"message\\\":\\\" not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z]\\\\nI1010 13:19:19.610281 6728 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1010 13:19:19.610285 6728 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 13:19:19.610282 6728 obj_retry.go:285] Attempting retry of *v1.Pod openshift-multus/network-metrics-daemon-xgh8p before timer (time: 2025-10-10 13:19:20.777105376 +0000 UTC m=+1.753565904): skip\\\\nI1010 13:19:19.610292 6728 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1010 13:19:19.610295 6728 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1010 13:19:19.610297 6728 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 13:19:19.610300 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:19:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:26Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.924845 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:26Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.943043 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:26Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.956646 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:26Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.965223 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:26Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.973915 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87b5334e-a097-425c-9a79-e62f4bed1f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81717c65162b43bab3af2ecbae6ec4e49ab3e89c9631187731fe36770d4c97c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701af5e4ef1c9e8c4f58ef5222fb6aee23624a16575e5b307ec14672c4143bff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://701af5e4ef1c9e8c4f58ef5222fb6aee23624a16575e5b307ec14672c4143bff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:26Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:26 crc kubenswrapper[4745]: I1010 13:19:26.986021 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:26Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:26.999888 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:26Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.013868 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:27Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.014011 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.014067 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.014084 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.014108 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.014126 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:27Z","lastTransitionTime":"2025-10-10T13:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.024401 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:27Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.116171 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.116202 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.116211 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.116224 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.116233 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:27Z","lastTransitionTime":"2025-10-10T13:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.217836 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.217863 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.217872 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.217884 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.217892 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:27Z","lastTransitionTime":"2025-10-10T13:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.319825 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.319874 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.319885 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.319905 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.319918 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:27Z","lastTransitionTime":"2025-10-10T13:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.424411 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.424454 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.424464 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.424478 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.424489 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:27Z","lastTransitionTime":"2025-10-10T13:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.527041 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.527545 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.527815 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.528148 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.528299 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:27Z","lastTransitionTime":"2025-10-10T13:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.631041 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.631126 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.631142 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.631164 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.631181 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:27Z","lastTransitionTime":"2025-10-10T13:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.734459 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.734506 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.734517 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.734534 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.734548 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:27Z","lastTransitionTime":"2025-10-10T13:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.744958 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.745082 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:27 crc kubenswrapper[4745]: E1010 13:19:27.745233 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:27 crc kubenswrapper[4745]: E1010 13:19:27.746042 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.837877 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.837940 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.837959 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.837980 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.837997 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:27Z","lastTransitionTime":"2025-10-10T13:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.941259 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.941309 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.941320 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.941338 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:27 crc kubenswrapper[4745]: I1010 13:19:27.941352 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:27Z","lastTransitionTime":"2025-10-10T13:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.044657 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.044797 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.044825 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.044856 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.044879 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:28Z","lastTransitionTime":"2025-10-10T13:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.151154 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.151229 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.151249 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.151275 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.151294 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:28Z","lastTransitionTime":"2025-10-10T13:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.254055 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.254118 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.254131 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.254152 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.254167 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:28Z","lastTransitionTime":"2025-10-10T13:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.356942 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.357019 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.357038 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.357067 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.357086 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:28Z","lastTransitionTime":"2025-10-10T13:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.459690 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.459813 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.459832 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.459856 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.459873 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:28Z","lastTransitionTime":"2025-10-10T13:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.563113 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.563166 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.563183 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.563206 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.563223 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:28Z","lastTransitionTime":"2025-10-10T13:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.666528 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.666960 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.667113 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.667300 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.667440 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:28Z","lastTransitionTime":"2025-10-10T13:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.744860 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.744952 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:28 crc kubenswrapper[4745]: E1010 13:19:28.745074 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:28 crc kubenswrapper[4745]: E1010 13:19:28.745202 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.770493 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.770542 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.770560 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.770583 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.770604 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:28Z","lastTransitionTime":"2025-10-10T13:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.873154 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.873224 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.873253 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.873284 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.873306 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:28Z","lastTransitionTime":"2025-10-10T13:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.976958 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.977033 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.977056 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.977091 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:28 crc kubenswrapper[4745]: I1010 13:19:28.977113 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:28Z","lastTransitionTime":"2025-10-10T13:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.080191 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.080261 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.080278 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.080303 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.080324 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:29Z","lastTransitionTime":"2025-10-10T13:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.184324 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.184378 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.184390 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.184405 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.184415 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:29Z","lastTransitionTime":"2025-10-10T13:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.286987 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.287037 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.287046 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.287060 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.287069 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:29Z","lastTransitionTime":"2025-10-10T13:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.389921 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.389993 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.390007 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.390029 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.390042 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:29Z","lastTransitionTime":"2025-10-10T13:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.493914 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.493999 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.494036 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.494067 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.494088 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:29Z","lastTransitionTime":"2025-10-10T13:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.597789 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.597868 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.597891 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.597922 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.597944 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:29Z","lastTransitionTime":"2025-10-10T13:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.701883 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.701977 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.702011 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.702043 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.702067 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:29Z","lastTransitionTime":"2025-10-10T13:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.744651 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.744716 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:29 crc kubenswrapper[4745]: E1010 13:19:29.744936 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:29 crc kubenswrapper[4745]: E1010 13:19:29.745138 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.805614 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.805689 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.805712 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.805774 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.805798 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:29Z","lastTransitionTime":"2025-10-10T13:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.910634 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.911017 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.911094 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.911131 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:29 crc kubenswrapper[4745]: I1010 13:19:29.911174 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:29Z","lastTransitionTime":"2025-10-10T13:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.014701 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.014847 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.014870 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.014923 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.014946 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:30Z","lastTransitionTime":"2025-10-10T13:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.077632 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:19:30 crc kubenswrapper[4745]: E1010 13:19:30.078039 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:34.077970264 +0000 UTC m=+147.975627067 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.118296 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.118380 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.118404 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.118437 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.118461 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:30Z","lastTransitionTime":"2025-10-10T13:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.179664 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.179726 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.179825 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.179900 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:30 crc kubenswrapper[4745]: E1010 13:19:30.180054 4745 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 13:19:30 crc kubenswrapper[4745]: E1010 13:19:30.180100 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 13:19:30 crc kubenswrapper[4745]: E1010 13:19:30.180099 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 13:19:30 crc kubenswrapper[4745]: E1010 13:19:30.180158 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 13:19:30 crc kubenswrapper[4745]: E1010 13:19:30.180174 4745 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 13:19:30 crc kubenswrapper[4745]: E1010 13:19:30.180229 4745 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:19:30 crc kubenswrapper[4745]: E1010 13:19:30.180148 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 13:20:34.180123076 +0000 UTC m=+148.077779889 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 13:19:30 crc kubenswrapper[4745]: E1010 13:19:30.180183 4745 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:19:30 crc kubenswrapper[4745]: E1010 13:19:30.180354 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 13:20:34.180302081 +0000 UTC m=+148.077958874 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:19:30 crc kubenswrapper[4745]: E1010 13:19:30.180455 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 13:20:34.180422874 +0000 UTC m=+148.078079677 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 13:19:30 crc kubenswrapper[4745]: E1010 13:19:30.180102 4745 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 13:19:30 crc kubenswrapper[4745]: E1010 13:19:30.180532 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 13:20:34.180513467 +0000 UTC m=+148.078170260 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.222224 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.222290 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.222330 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.222354 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.222371 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:30Z","lastTransitionTime":"2025-10-10T13:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.326129 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.326196 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.326217 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.326242 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.326259 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:30Z","lastTransitionTime":"2025-10-10T13:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.430095 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.430158 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.430174 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.430199 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.430217 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:30Z","lastTransitionTime":"2025-10-10T13:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.533236 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.533313 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.533332 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.533359 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.533378 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:30Z","lastTransitionTime":"2025-10-10T13:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.638195 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.638273 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.638292 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.638323 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.638344 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:30Z","lastTransitionTime":"2025-10-10T13:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.742465 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.742533 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.742552 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.742580 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.742599 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:30Z","lastTransitionTime":"2025-10-10T13:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.744374 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.744496 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:30 crc kubenswrapper[4745]: E1010 13:19:30.744666 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:30 crc kubenswrapper[4745]: E1010 13:19:30.745004 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.846871 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.846957 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.846977 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.847008 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.847029 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:30Z","lastTransitionTime":"2025-10-10T13:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.950151 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.950237 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.950256 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.950287 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:30 crc kubenswrapper[4745]: I1010 13:19:30.950307 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:30Z","lastTransitionTime":"2025-10-10T13:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.053578 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.053638 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.053694 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.053717 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.053757 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:31Z","lastTransitionTime":"2025-10-10T13:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.158502 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.158571 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.158596 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.158630 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.158651 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:31Z","lastTransitionTime":"2025-10-10T13:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.261959 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.262023 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.262041 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.262281 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.262303 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:31Z","lastTransitionTime":"2025-10-10T13:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.365970 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.366048 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.366070 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.366102 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.366129 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:31Z","lastTransitionTime":"2025-10-10T13:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.470326 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.470401 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.470423 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.470446 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.470465 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:31Z","lastTransitionTime":"2025-10-10T13:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.574231 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.574299 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.574315 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.574343 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.574360 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:31Z","lastTransitionTime":"2025-10-10T13:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.678012 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.678076 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.678094 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.678124 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.678145 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:31Z","lastTransitionTime":"2025-10-10T13:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.744307 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.744399 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:31 crc kubenswrapper[4745]: E1010 13:19:31.744509 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:31 crc kubenswrapper[4745]: E1010 13:19:31.744751 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.782301 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.782380 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.782522 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.782607 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.782629 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:31Z","lastTransitionTime":"2025-10-10T13:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.886368 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.886923 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.887101 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.887262 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.887400 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:31Z","lastTransitionTime":"2025-10-10T13:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.990674 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.991214 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.991232 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.991257 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:31 crc kubenswrapper[4745]: I1010 13:19:31.991274 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:31Z","lastTransitionTime":"2025-10-10T13:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.094277 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.094333 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.094349 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.094381 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.094398 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:32Z","lastTransitionTime":"2025-10-10T13:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.198264 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.198329 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.198348 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.198372 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.198389 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:32Z","lastTransitionTime":"2025-10-10T13:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.300911 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.300997 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.301018 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.301036 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.301047 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:32Z","lastTransitionTime":"2025-10-10T13:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.403844 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.403899 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.403911 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.403927 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.403939 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:32Z","lastTransitionTime":"2025-10-10T13:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.507126 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.507179 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.507195 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.507218 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.507234 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:32Z","lastTransitionTime":"2025-10-10T13:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.610174 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.610227 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.610245 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.610268 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.610284 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:32Z","lastTransitionTime":"2025-10-10T13:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.713531 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.713608 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.713628 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.713654 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.713674 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:32Z","lastTransitionTime":"2025-10-10T13:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.744458 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.744511 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:32 crc kubenswrapper[4745]: E1010 13:19:32.744664 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:32 crc kubenswrapper[4745]: E1010 13:19:32.744816 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.817069 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.817180 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.817199 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.817225 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.817245 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:32Z","lastTransitionTime":"2025-10-10T13:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.921221 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.921291 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.921309 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.921341 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:32 crc kubenswrapper[4745]: I1010 13:19:32.921362 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:32Z","lastTransitionTime":"2025-10-10T13:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.025686 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.026310 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.026460 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.026640 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.026820 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:33Z","lastTransitionTime":"2025-10-10T13:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.135567 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.136066 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.136229 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.136435 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.136598 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:33Z","lastTransitionTime":"2025-10-10T13:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.240781 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.240857 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.240881 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.240912 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.240935 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:33Z","lastTransitionTime":"2025-10-10T13:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.344407 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.344535 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.344558 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.344622 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.344643 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:33Z","lastTransitionTime":"2025-10-10T13:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.447209 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.447299 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.447324 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.447353 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.447372 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:33Z","lastTransitionTime":"2025-10-10T13:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.551017 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.551095 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.551113 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.551141 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.551158 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:33Z","lastTransitionTime":"2025-10-10T13:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.654813 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.654889 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.654912 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.654942 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.654959 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:33Z","lastTransitionTime":"2025-10-10T13:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.744489 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.744549 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:33 crc kubenswrapper[4745]: E1010 13:19:33.744713 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:33 crc kubenswrapper[4745]: E1010 13:19:33.744867 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.747310 4745 scope.go:117] "RemoveContainer" containerID="b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8" Oct 10 13:19:33 crc kubenswrapper[4745]: E1010 13:19:33.747868 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podUID="687bad4e-baff-4e09-bb33-def12c499092" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.757865 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.757943 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.757997 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.758023 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.758041 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:33Z","lastTransitionTime":"2025-10-10T13:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.861532 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.861594 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.861611 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.861638 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.861659 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:33Z","lastTransitionTime":"2025-10-10T13:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.965247 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.965317 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.965340 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.965370 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:33 crc kubenswrapper[4745]: I1010 13:19:33.965395 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:33Z","lastTransitionTime":"2025-10-10T13:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.069614 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.069675 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.069694 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.069721 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.069769 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:34Z","lastTransitionTime":"2025-10-10T13:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.172634 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.172741 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.172756 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.172778 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.172792 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:34Z","lastTransitionTime":"2025-10-10T13:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.274477 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.274597 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.274608 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.274624 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.274636 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:34Z","lastTransitionTime":"2025-10-10T13:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.377406 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.377445 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.377454 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.377469 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.377478 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:34Z","lastTransitionTime":"2025-10-10T13:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.480977 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.481061 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.481084 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.481117 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.481141 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:34Z","lastTransitionTime":"2025-10-10T13:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.584362 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.584438 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.584456 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.584485 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.584509 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:34Z","lastTransitionTime":"2025-10-10T13:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.687975 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.688056 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.688074 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.688099 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.688116 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:34Z","lastTransitionTime":"2025-10-10T13:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.744319 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:34 crc kubenswrapper[4745]: E1010 13:19:34.744721 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.745172 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:34 crc kubenswrapper[4745]: E1010 13:19:34.745382 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.790808 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.790870 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.790893 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.790927 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.790952 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:34Z","lastTransitionTime":"2025-10-10T13:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.824048 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.824111 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.824123 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.824144 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.824157 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:34Z","lastTransitionTime":"2025-10-10T13:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:34 crc kubenswrapper[4745]: E1010 13:19:34.840117 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.845206 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.845270 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.845288 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.845320 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.845340 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:34Z","lastTransitionTime":"2025-10-10T13:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:34 crc kubenswrapper[4745]: E1010 13:19:34.860351 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.865810 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.865883 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.865902 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.865929 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.865948 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:34Z","lastTransitionTime":"2025-10-10T13:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:34 crc kubenswrapper[4745]: E1010 13:19:34.884559 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.889343 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.889407 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.889431 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.889460 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.889482 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:34Z","lastTransitionTime":"2025-10-10T13:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:34 crc kubenswrapper[4745]: E1010 13:19:34.914680 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.920014 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.920066 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.920083 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.920107 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.920123 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:34Z","lastTransitionTime":"2025-10-10T13:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:34 crc kubenswrapper[4745]: E1010 13:19:34.939918 4745 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53df7a22-5107-4004-bd75-de252c4adb13\\\",\\\"systemUUID\\\":\\\"8fd7c1b2-971d-45b1-b219-6a04ef38b544\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:34Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:34 crc kubenswrapper[4745]: E1010 13:19:34.940137 4745 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.942773 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.942810 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.942825 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.942846 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:34 crc kubenswrapper[4745]: I1010 13:19:34.942860 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:34Z","lastTransitionTime":"2025-10-10T13:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.046000 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.046066 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.046084 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.046108 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.046125 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:35Z","lastTransitionTime":"2025-10-10T13:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.148898 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.148949 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.148959 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.148976 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.148987 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:35Z","lastTransitionTime":"2025-10-10T13:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.252077 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.252179 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.252213 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.252250 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.252278 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:35Z","lastTransitionTime":"2025-10-10T13:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.355255 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.355324 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.355351 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.355383 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.355406 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:35Z","lastTransitionTime":"2025-10-10T13:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.458569 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.458647 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.458667 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.458691 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.458708 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:35Z","lastTransitionTime":"2025-10-10T13:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.561370 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.561438 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.561461 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.561488 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.561509 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:35Z","lastTransitionTime":"2025-10-10T13:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.664163 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.664221 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.664239 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.664261 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.664278 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:35Z","lastTransitionTime":"2025-10-10T13:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.744561 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.744563 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:35 crc kubenswrapper[4745]: E1010 13:19:35.744787 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:35 crc kubenswrapper[4745]: E1010 13:19:35.745012 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.766901 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.766957 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.766979 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.767004 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.767025 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:35Z","lastTransitionTime":"2025-10-10T13:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.869843 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.869911 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.869939 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.869968 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.869988 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:35Z","lastTransitionTime":"2025-10-10T13:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.973454 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.973531 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.973549 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.973575 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:35 crc kubenswrapper[4745]: I1010 13:19:35.973593 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:35Z","lastTransitionTime":"2025-10-10T13:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.076626 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.076680 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.076696 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.076721 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.076814 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:36Z","lastTransitionTime":"2025-10-10T13:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.179432 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.179488 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.179504 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.179525 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.179540 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:36Z","lastTransitionTime":"2025-10-10T13:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.281622 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.281692 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.281709 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.281762 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.281794 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:36Z","lastTransitionTime":"2025-10-10T13:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.384103 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.384168 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.384186 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.384213 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.384233 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:36Z","lastTransitionTime":"2025-10-10T13:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.488282 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.488356 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.488381 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.488416 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.488441 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:36Z","lastTransitionTime":"2025-10-10T13:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.591204 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.591252 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.591263 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.591283 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.591296 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:36Z","lastTransitionTime":"2025-10-10T13:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.694478 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.694513 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.694523 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.694541 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.694551 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:36Z","lastTransitionTime":"2025-10-10T13:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.744863 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.744883 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:36 crc kubenswrapper[4745]: E1010 13:19:36.745072 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:36 crc kubenswrapper[4745]: E1010 13:19:36.745178 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.766195 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.785561 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rdsn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bce814bc-030f-402e-ae0a-546f1c63edbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda2ad46a4264696356d8dac1e33c4673bc57ff767831db38fb28cf49e5b2981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9ltc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rdsn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.797931 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.797996 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.798015 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.798044 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.798068 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:36Z","lastTransitionTime":"2025-10-10T13:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.802521 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87b5334e-a097-425c-9a79-e62f4bed1f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81717c65162b43bab3af2ecbae6ec4e49ab3e89c9631187731fe36770d4c97c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701af5e4ef1c9e8c4f58ef5222fb6aee23624a16575e5b307ec14672c4143bff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://701af5e4ef1c9e8c4f58ef5222fb6aee23624a16575e5b307ec14672c4143bff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.822055 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.847444 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlths" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89f8c1b-e9aa-48a9-8072-254db014f8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35242a9c41e606b07142e5ed5e6f0c2052bfcd64cd2ecc8a861c4fd49a9a7a6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b627e88975f53a8be4baac0c64c476858d54bb1bb74c4f068f37a82bcb21e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57336633f5757e6030e6a08f110c42682e79885cfbb7566cdb73bbf82dc7d7d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ffdf64c4c807e396fd492f747a691be8df1644b47694937c0044c1ab354447\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fc6fd305557fd06e3ccbbad0c0c27f2e832834ee43b2ef4ff504ebfe49dba95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdeed12e0d37260e070796efe4fd67a59bf4f88f19a0629f88f66fb737f5eada\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f76476af71a16e6753e7d86c87f50c2343e27e92e11404dfbd1cc3e8c70e029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6btr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlths\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.863677 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f433c55c51fdcf99076b1bb329d9905c0d3bf2e83e08051de7de877f2abf8171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.891083 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.901887 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.901945 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.901957 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.901976 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.901990 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:36Z","lastTransitionTime":"2025-10-10T13:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.910881 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-27d2n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:19:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f78611b429d8411902427cc4f7dda27fb9f1ea97db3b312737d4bbc38d350a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:19:15Z\\\",\\\"message\\\":\\\"2025-10-10T13:18:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e6a3b6a9-97e0-4aad-9ab9-eb4f852eb8d1\\\\n2025-10-10T13:18:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e6a3b6a9-97e0-4aad-9ab9-eb4f852eb8d1 to /host/opt/cni/bin/\\\\n2025-10-10T13:18:30Z [verbose] multus-daemon started\\\\n2025-10-10T13:18:30Z [verbose] Readiness Indicator file check\\\\n2025-10-10T13:19:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:19:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sg98g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-27d2n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.927424 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b60d4139-055c-4d5a-9529-1c81f5639053\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d60840951da17cbaf07c1012a910b887b010aa3daaa8c44f15529ab0fb6eb113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba8f5784fc2930cd919753d4475aed17e79903724267ccc3ee6fa41f9510f7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stvrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62rnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.940891 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82c4e63a-009f-4720-87b5-640a8ec65117\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrj8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xgh8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.958459 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923c385c-8102-4892-a4a3-67b6408545ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://193833f7e032a32aae576702c6965f7ea2869114b86d1a0349e7ad787869c74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab4465f351419e32376cc0b2062225c6faac1e27dce51bcf3c1ca4bb590fc4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f6659c3fefc2b4892b34a407a26418aa86c780a7e03c9120794f4396c08913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45772aef1a4894eefa3e2a3b9dcf9cca7712e08d622a65a9b372e715a7a7b407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.975495 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f41db15410c59cebba15d0a40813ecc06923b6f3307d8e95744168b370d7d3e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftf65\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dnq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:36 crc kubenswrapper[4745]: I1010 13:19:36.996475 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"085cbb09-d34c-43c2-a1d3-37a0365d8871\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d117ca5e53f8e25611f442db311e11d45c0e1b4f5ce4cb6125b99f2f7da94dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10eedd7130dd7cf0fa52daadcea2bc96030548807714e693858bb648ba40c3f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a6b25869a83781e28717239ae5686705c3b44a61c120b844c26069517cba1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abce00b9f6abce0553f024cb841e8027d75a261ffe35e77f6499c446b2b3bc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72d4e7edc31dd887afdc88ce91bb9079d1d71d4b0cbfcd013a0b9f79ef4b4093\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 13:18:20.262231 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 13:18:20.263675 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1206230131/tls.crt::/tmp/serving-cert-1206230131/tls.key\\\\\\\"\\\\nI1010 13:18:26.222787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 13:18:26.226529 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 13:18:26.226550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 13:18:26.226573 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 13:18:26.226578 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 13:18:26.232721 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 13:18:26.232967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232972 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 13:18:26.232977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 13:18:26.232980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 13:18:26.232982 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 13:18:26.232985 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 13:18:26.232871 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 13:18:26.236972 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://227f96ef6b7a16e25e633f408c128163781ec04eed1779629d6baf0e18070ae4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f574bd89107f302c3c713fec5381d217e2e926f30c9d6c3930f20f81523e35fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:36Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.006193 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.006259 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.006269 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.006307 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.006324 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:37Z","lastTransitionTime":"2025-10-10T13:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.017516 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fea6eb539e62efe23d9dccb3e2471daebb4e5cc4f97d298b1be6fad6dbcdd0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:37Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.037716 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cade5b2fa0000255ad8f017a9a3a557e2c32bedcd9d62919ef0b2c333b77db2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f898df5dd8d5c2c9f47e7934407de13a5fce1906177d198b3bb129759f18c791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:37Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.054234 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tzws7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c112c93-ee29-4baa-aa3c-e370f4492bf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b373ec3090439da29a106da2f879b44a08fb49b8711c31ff329edda7921168cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn4tw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tzws7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:37Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.086432 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687bad4e-baff-4e09-bb33-def12c499092\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T13:19:19Z\\\",\\\"message\\\":\\\" not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:19Z is after 2025-08-24T17:21:41Z]\\\\nI1010 13:19:19.610281 6728 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1010 13:19:19.610285 6728 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 13:19:19.610282 6728 obj_retry.go:285] Attempting retry of *v1.Pod openshift-multus/network-metrics-daemon-xgh8p before timer (time: 2025-10-10 13:19:20.777105376 +0000 UTC m=+1.753565904): skip\\\\nI1010 13:19:19.610292 6728 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1010 13:19:19.610295 6728 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1010 13:19:19.610297 6728 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 13:19:19.610300 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T13:19:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9sbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5bh9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:37Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.106601 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335ab7fc-59cb-4686-b401-eedf781d39f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52dbc7138dc491632429a0c71ea380bdde6ddea95b3f33d89e57999e8015d771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af7fc4f98c02522a0f6287a5e3890c81030703020d34f2c7d00e23b33ac4862b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf67ab66a1f63eeab02ff26684abacefbb7ce183d0038b6449ad987ee0fc70fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c445f1339fbdbd8256e878214bbc7a6f7f458f04a2e7bf0fbe71471af224cf08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:37Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.110160 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.110217 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.110264 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.110289 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.110307 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:37Z","lastTransitionTime":"2025-10-10T13:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.134720 4745 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aceaad01-0e34-45c7-a7bc-3a26d9b2056d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T13:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b5d97f24bbbf1f1119ff235587058175f0b3cb278db096a2a24eb17a1554539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d4e87b997eff791a33da6d76d86b4939e0ed4c38963a22c7c53e9c33f6f90cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bcfbe34f939a5b6cfd94e52909a29b124128bcfe44d9f7cfa38bd2795c4cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a9db0b82e7cee4991e990c3d7aec4faeab5d6232ac496fd898ebc7adc7bb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5882ffc86652907fc5e2caf4a808c0e8c323cf5ed4e883887cba78e3e885d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T13:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f9ae17205da4410885ae72b666aaf681beecdd709edf17d348d6b47683c8a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78026a2a6391248bceade8e301d191447e0b1b3ce27640ff7085c49d94d70579\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2eec59ef12528db0ed5aecb84540708ae78d4f77ce8deb0db3c361c7a2ac4f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T13:18:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T13:18:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T13:18:06Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T13:19:37Z is after 2025-08-24T17:21:41Z" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.213629 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.213686 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.213705 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.213783 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.213813 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:37Z","lastTransitionTime":"2025-10-10T13:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.317325 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.317394 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.317418 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.317451 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.317476 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:37Z","lastTransitionTime":"2025-10-10T13:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.419966 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.420009 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.420021 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.420037 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.420050 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:37Z","lastTransitionTime":"2025-10-10T13:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.523570 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.523654 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.523675 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.523710 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.523776 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:37Z","lastTransitionTime":"2025-10-10T13:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.627151 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.627215 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.627227 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.627250 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.627271 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:37Z","lastTransitionTime":"2025-10-10T13:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.731399 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.731444 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.731454 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.731473 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.731486 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:37Z","lastTransitionTime":"2025-10-10T13:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.744581 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.744590 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:37 crc kubenswrapper[4745]: E1010 13:19:37.745098 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:37 crc kubenswrapper[4745]: E1010 13:19:37.745271 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.835446 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.835525 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.835554 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.835590 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.835614 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:37Z","lastTransitionTime":"2025-10-10T13:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.938985 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.939073 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.939096 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.939128 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:37 crc kubenswrapper[4745]: I1010 13:19:37.939150 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:37Z","lastTransitionTime":"2025-10-10T13:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.043052 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.043129 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.043146 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.043177 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.043197 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:38Z","lastTransitionTime":"2025-10-10T13:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.147176 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.147250 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.147269 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.147296 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.147318 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:38Z","lastTransitionTime":"2025-10-10T13:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.250621 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.250707 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.250774 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.250812 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.250834 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:38Z","lastTransitionTime":"2025-10-10T13:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.354787 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.354853 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.354872 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.354899 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.354917 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:38Z","lastTransitionTime":"2025-10-10T13:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.458860 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.458943 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.458963 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.458990 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.459008 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:38Z","lastTransitionTime":"2025-10-10T13:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.561869 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.561944 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.561960 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.561986 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.562006 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:38Z","lastTransitionTime":"2025-10-10T13:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.665176 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.665262 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.665280 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.665311 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.665401 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:38Z","lastTransitionTime":"2025-10-10T13:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.744460 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.744632 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:38 crc kubenswrapper[4745]: E1010 13:19:38.744777 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:38 crc kubenswrapper[4745]: E1010 13:19:38.744911 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.769245 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.769291 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.769301 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.769317 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.769328 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:38Z","lastTransitionTime":"2025-10-10T13:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.872540 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.872622 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.872642 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.872675 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.872697 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:38Z","lastTransitionTime":"2025-10-10T13:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.976825 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.976887 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.976909 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.976940 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:38 crc kubenswrapper[4745]: I1010 13:19:38.976965 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:38Z","lastTransitionTime":"2025-10-10T13:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.080999 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.081082 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.081104 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.081140 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.081163 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:39Z","lastTransitionTime":"2025-10-10T13:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.185457 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.185539 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.185559 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.185592 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.185611 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:39Z","lastTransitionTime":"2025-10-10T13:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.289337 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.289401 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.289421 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.289447 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.289467 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:39Z","lastTransitionTime":"2025-10-10T13:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.392972 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.393046 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.393065 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.393098 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.393119 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:39Z","lastTransitionTime":"2025-10-10T13:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.497807 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.497881 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.497896 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.497922 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.497939 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:39Z","lastTransitionTime":"2025-10-10T13:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.600862 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.600910 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.600949 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.600975 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.600990 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:39Z","lastTransitionTime":"2025-10-10T13:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.704485 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.704552 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.704569 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.704596 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.704616 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:39Z","lastTransitionTime":"2025-10-10T13:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.744349 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.744362 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:39 crc kubenswrapper[4745]: E1010 13:19:39.744545 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:39 crc kubenswrapper[4745]: E1010 13:19:39.744663 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.808116 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.808167 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.808178 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.808196 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.808208 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:39Z","lastTransitionTime":"2025-10-10T13:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.911164 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.911281 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.911339 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.911369 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:39 crc kubenswrapper[4745]: I1010 13:19:39.911388 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:39Z","lastTransitionTime":"2025-10-10T13:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.014845 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.014938 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.014960 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.014995 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.015020 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:40Z","lastTransitionTime":"2025-10-10T13:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.119238 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.119342 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.119367 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.119403 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.119428 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:40Z","lastTransitionTime":"2025-10-10T13:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.223292 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.223393 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.223418 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.223451 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.223476 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:40Z","lastTransitionTime":"2025-10-10T13:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.328532 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.328614 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.328633 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.328661 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.328685 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:40Z","lastTransitionTime":"2025-10-10T13:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.432992 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.433069 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.433088 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.433117 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.433140 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:40Z","lastTransitionTime":"2025-10-10T13:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.534958 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.535021 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.535037 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.535058 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.535074 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:40Z","lastTransitionTime":"2025-10-10T13:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.637761 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.637819 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.637835 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.637859 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.637877 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:40Z","lastTransitionTime":"2025-10-10T13:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.741472 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.741579 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.741603 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.741632 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.741655 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:40Z","lastTransitionTime":"2025-10-10T13:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.745120 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.745305 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:40 crc kubenswrapper[4745]: E1010 13:19:40.745414 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:40 crc kubenswrapper[4745]: E1010 13:19:40.745589 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.844665 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.844765 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.844783 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.844805 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.844823 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:40Z","lastTransitionTime":"2025-10-10T13:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.947364 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.947424 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.947440 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.947465 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:40 crc kubenswrapper[4745]: I1010 13:19:40.947482 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:40Z","lastTransitionTime":"2025-10-10T13:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.051027 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.051094 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.051114 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.051139 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.051158 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:41Z","lastTransitionTime":"2025-10-10T13:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.153951 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.154024 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.154043 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.154074 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.154093 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:41Z","lastTransitionTime":"2025-10-10T13:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.256371 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.256445 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.256468 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.256495 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.256514 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:41Z","lastTransitionTime":"2025-10-10T13:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.359059 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.359138 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.359162 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.359192 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.359218 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:41Z","lastTransitionTime":"2025-10-10T13:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.463308 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.463395 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.463419 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.463453 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.463476 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:41Z","lastTransitionTime":"2025-10-10T13:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.566507 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.566568 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.566584 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.566609 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.566626 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:41Z","lastTransitionTime":"2025-10-10T13:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.670364 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.670446 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.670466 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.670483 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.670494 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:41Z","lastTransitionTime":"2025-10-10T13:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.744910 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.744962 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:41 crc kubenswrapper[4745]: E1010 13:19:41.745143 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:41 crc kubenswrapper[4745]: E1010 13:19:41.745328 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.773825 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.773888 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.773907 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.773929 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.773946 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:41Z","lastTransitionTime":"2025-10-10T13:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.876425 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.876522 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.876574 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.876602 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.876621 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:41Z","lastTransitionTime":"2025-10-10T13:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.979606 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.979682 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.979704 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.979781 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:41 crc kubenswrapper[4745]: I1010 13:19:41.979809 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:41Z","lastTransitionTime":"2025-10-10T13:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.083139 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.083209 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.083228 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.083252 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.083272 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:42Z","lastTransitionTime":"2025-10-10T13:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.186893 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.186989 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.187016 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.187055 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.187087 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:42Z","lastTransitionTime":"2025-10-10T13:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.290322 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.290396 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.290416 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.290446 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.290468 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:42Z","lastTransitionTime":"2025-10-10T13:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.394563 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.394641 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.394664 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.394696 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.394716 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:42Z","lastTransitionTime":"2025-10-10T13:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.498247 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.498334 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.498354 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.498380 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.498427 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:42Z","lastTransitionTime":"2025-10-10T13:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.603554 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.603612 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.603624 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.603649 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.603661 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:42Z","lastTransitionTime":"2025-10-10T13:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.707053 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.707103 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.707151 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.707176 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.707194 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:42Z","lastTransitionTime":"2025-10-10T13:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.744651 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.744716 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:42 crc kubenswrapper[4745]: E1010 13:19:42.744883 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:42 crc kubenswrapper[4745]: E1010 13:19:42.745405 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.810786 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.810849 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.810868 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.810892 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.810910 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:42Z","lastTransitionTime":"2025-10-10T13:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.915095 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.915170 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.915198 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.915231 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:42 crc kubenswrapper[4745]: I1010 13:19:42.915256 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:42Z","lastTransitionTime":"2025-10-10T13:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.018761 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.018812 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.018829 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.018851 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.018869 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:43Z","lastTransitionTime":"2025-10-10T13:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.122629 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.122673 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.122686 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.122705 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.122719 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:43Z","lastTransitionTime":"2025-10-10T13:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.225218 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.225609 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.225783 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.225918 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.226036 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:43Z","lastTransitionTime":"2025-10-10T13:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.329155 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.329225 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.329246 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.329269 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.329287 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:43Z","lastTransitionTime":"2025-10-10T13:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.432687 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.432805 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.432824 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.432848 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.432865 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:43Z","lastTransitionTime":"2025-10-10T13:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.536462 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.536545 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.536610 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.536642 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.536665 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:43Z","lastTransitionTime":"2025-10-10T13:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.639888 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.639944 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.639960 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.639982 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.639997 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:43Z","lastTransitionTime":"2025-10-10T13:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.742447 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.742510 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.742535 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.742561 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.742583 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:43Z","lastTransitionTime":"2025-10-10T13:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.743942 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.743974 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:43 crc kubenswrapper[4745]: E1010 13:19:43.744126 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:43 crc kubenswrapper[4745]: E1010 13:19:43.744408 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.845868 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.845930 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.845948 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.845971 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.845989 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:43Z","lastTransitionTime":"2025-10-10T13:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.949016 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.949116 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.949139 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.949166 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:43 crc kubenswrapper[4745]: I1010 13:19:43.949188 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:43Z","lastTransitionTime":"2025-10-10T13:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.052900 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.052982 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.053003 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.053029 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.053046 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:44Z","lastTransitionTime":"2025-10-10T13:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.156877 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.156975 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.156997 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.157022 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.157039 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:44Z","lastTransitionTime":"2025-10-10T13:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.260690 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.260784 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.260802 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.260828 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.260846 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:44Z","lastTransitionTime":"2025-10-10T13:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.363950 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.364006 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.364018 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.364125 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.364139 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:44Z","lastTransitionTime":"2025-10-10T13:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.467469 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.467538 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.467568 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.467592 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.467612 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:44Z","lastTransitionTime":"2025-10-10T13:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.570615 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.570654 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.570664 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.570679 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.570691 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:44Z","lastTransitionTime":"2025-10-10T13:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.673393 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.673446 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.673461 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.673481 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.673495 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:44Z","lastTransitionTime":"2025-10-10T13:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.745039 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:44 crc kubenswrapper[4745]: E1010 13:19:44.745237 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.745703 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:44 crc kubenswrapper[4745]: E1010 13:19:44.745891 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.776667 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.776766 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.776790 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.776815 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.776857 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:44Z","lastTransitionTime":"2025-10-10T13:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.880280 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.880351 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.880369 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.880395 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.880414 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:44Z","lastTransitionTime":"2025-10-10T13:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.984247 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.984317 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.984336 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.984363 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:44 crc kubenswrapper[4745]: I1010 13:19:44.984381 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:44Z","lastTransitionTime":"2025-10-10T13:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.086965 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.087125 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.087145 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.087217 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.087237 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:45Z","lastTransitionTime":"2025-10-10T13:19:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.132367 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.132431 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.132448 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.132474 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.132491 4745 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T13:19:45Z","lastTransitionTime":"2025-10-10T13:19:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.200615 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg"] Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.201827 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.204218 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.204287 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.207281 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.207607 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.278991 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=76.278976355 podStartE2EDuration="1m16.278976355s" podCreationTimestamp="2025-10-10 13:18:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:19:45.250049452 +0000 UTC m=+99.147706215" watchObservedRunningTime="2025-10-10 13:19:45.278976355 +0000 UTC m=+99.176633118" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.279480 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=81.279476439 podStartE2EDuration="1m21.279476439s" podCreationTimestamp="2025-10-10 13:18:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:19:45.278697147 +0000 UTC m=+99.176353900" watchObservedRunningTime="2025-10-10 13:19:45.279476439 +0000 UTC m=+99.177133202" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.289080 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2755a074-c254-4a9b-9a3a-63399d775510-service-ca\") pod \"cluster-version-operator-5c965bbfc6-m78fg\" (UID: \"2755a074-c254-4a9b-9a3a-63399d775510\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.289215 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2755a074-c254-4a9b-9a3a-63399d775510-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-m78fg\" (UID: \"2755a074-c254-4a9b-9a3a-63399d775510\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.289322 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2755a074-c254-4a9b-9a3a-63399d775510-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-m78fg\" (UID: \"2755a074-c254-4a9b-9a3a-63399d775510\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.289411 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2755a074-c254-4a9b-9a3a-63399d775510-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-m78fg\" (UID: \"2755a074-c254-4a9b-9a3a-63399d775510\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.289500 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2755a074-c254-4a9b-9a3a-63399d775510-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-m78fg\" (UID: \"2755a074-c254-4a9b-9a3a-63399d775510\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.320295 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=79.320278876 podStartE2EDuration="1m19.320278876s" podCreationTimestamp="2025-10-10 13:18:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:19:45.303393541 +0000 UTC m=+99.201050304" watchObservedRunningTime="2025-10-10 13:19:45.320278876 +0000 UTC m=+99.217935629" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.355599 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-tzws7" podStartSLOduration=78.355582678 podStartE2EDuration="1m18.355582678s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:19:45.355216988 +0000 UTC m=+99.252873751" watchObservedRunningTime="2025-10-10 13:19:45.355582678 +0000 UTC m=+99.253239441" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.390746 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2755a074-c254-4a9b-9a3a-63399d775510-service-ca\") pod \"cluster-version-operator-5c965bbfc6-m78fg\" (UID: \"2755a074-c254-4a9b-9a3a-63399d775510\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.391247 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2755a074-c254-4a9b-9a3a-63399d775510-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-m78fg\" (UID: \"2755a074-c254-4a9b-9a3a-63399d775510\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.391324 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2755a074-c254-4a9b-9a3a-63399d775510-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-m78fg\" (UID: \"2755a074-c254-4a9b-9a3a-63399d775510\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.391404 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2755a074-c254-4a9b-9a3a-63399d775510-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-m78fg\" (UID: \"2755a074-c254-4a9b-9a3a-63399d775510\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.391522 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2755a074-c254-4a9b-9a3a-63399d775510-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-m78fg\" (UID: \"2755a074-c254-4a9b-9a3a-63399d775510\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.391551 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2755a074-c254-4a9b-9a3a-63399d775510-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-m78fg\" (UID: \"2755a074-c254-4a9b-9a3a-63399d775510\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.391652 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2755a074-c254-4a9b-9a3a-63399d775510-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-m78fg\" (UID: \"2755a074-c254-4a9b-9a3a-63399d775510\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.393127 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2755a074-c254-4a9b-9a3a-63399d775510-service-ca\") pod \"cluster-version-operator-5c965bbfc6-m78fg\" (UID: \"2755a074-c254-4a9b-9a3a-63399d775510\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.404932 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2755a074-c254-4a9b-9a3a-63399d775510-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-m78fg\" (UID: \"2755a074-c254-4a9b-9a3a-63399d775510\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.406784 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=24.406716906 podStartE2EDuration="24.406716906s" podCreationTimestamp="2025-10-10 13:19:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:19:45.406181811 +0000 UTC m=+99.303838574" watchObservedRunningTime="2025-10-10 13:19:45.406716906 +0000 UTC m=+99.304373719" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.424316 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2755a074-c254-4a9b-9a3a-63399d775510-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-m78fg\" (UID: \"2755a074-c254-4a9b-9a3a-63399d775510\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.452665 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-7rdsn" podStartSLOduration=78.452640917 podStartE2EDuration="1m18.452640917s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:19:45.452308687 +0000 UTC m=+99.349965490" watchObservedRunningTime="2025-10-10 13:19:45.452640917 +0000 UTC m=+99.350297680" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.502488 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-rlths" podStartSLOduration=78.502469068 podStartE2EDuration="1m18.502469068s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:19:45.501567122 +0000 UTC m=+99.399223885" watchObservedRunningTime="2025-10-10 13:19:45.502469068 +0000 UTC m=+99.400125831" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.525008 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.537595 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=47.537565564 podStartE2EDuration="47.537565564s" podCreationTimestamp="2025-10-10 13:18:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:19:45.521791741 +0000 UTC m=+99.419448504" watchObservedRunningTime="2025-10-10 13:19:45.537565564 +0000 UTC m=+99.435222367" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.537996 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podStartSLOduration=78.537985296 podStartE2EDuration="1m18.537985296s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:19:45.537557714 +0000 UTC m=+99.435214537" watchObservedRunningTime="2025-10-10 13:19:45.537985296 +0000 UTC m=+99.435642099" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.578486 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-27d2n" podStartSLOduration=78.578470224 podStartE2EDuration="1m18.578470224s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:19:45.558635436 +0000 UTC m=+99.456292199" watchObservedRunningTime="2025-10-10 13:19:45.578470224 +0000 UTC m=+99.476126987" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.598544 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62rnv" podStartSLOduration=77.598527868 podStartE2EDuration="1m17.598527868s" podCreationTimestamp="2025-10-10 13:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:19:45.578866945 +0000 UTC m=+99.476523708" watchObservedRunningTime="2025-10-10 13:19:45.598527868 +0000 UTC m=+99.496184631" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.744404 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.744443 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:45 crc kubenswrapper[4745]: E1010 13:19:45.745064 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:45 crc kubenswrapper[4745]: E1010 13:19:45.745208 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:45 crc kubenswrapper[4745]: I1010 13:19:45.745888 4745 scope.go:117] "RemoveContainer" containerID="b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8" Oct 10 13:19:45 crc kubenswrapper[4745]: E1010 13:19:45.746154 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5bh9m_openshift-ovn-kubernetes(687bad4e-baff-4e09-bb33-def12c499092)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podUID="687bad4e-baff-4e09-bb33-def12c499092" Oct 10 13:19:46 crc kubenswrapper[4745]: I1010 13:19:46.099869 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs\") pod \"network-metrics-daemon-xgh8p\" (UID: \"82c4e63a-009f-4720-87b5-640a8ec65117\") " pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:46 crc kubenswrapper[4745]: E1010 13:19:46.099985 4745 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 13:19:46 crc kubenswrapper[4745]: E1010 13:19:46.100037 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs podName:82c4e63a-009f-4720-87b5-640a8ec65117 nodeName:}" failed. No retries permitted until 2025-10-10 13:20:50.100023336 +0000 UTC m=+163.997680099 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs") pod "network-metrics-daemon-xgh8p" (UID: "82c4e63a-009f-4720-87b5-640a8ec65117") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 13:19:46 crc kubenswrapper[4745]: I1010 13:19:46.317486 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" event={"ID":"2755a074-c254-4a9b-9a3a-63399d775510","Type":"ContainerStarted","Data":"4abcd04313f298b63d5a346fd16d241402aa07596e77edf742806056a7b7a8f9"} Oct 10 13:19:46 crc kubenswrapper[4745]: I1010 13:19:46.317565 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" event={"ID":"2755a074-c254-4a9b-9a3a-63399d775510","Type":"ContainerStarted","Data":"2e7abfec81fc53bd74d6e81bcc6fc185f2a67640f2d4bb64b703f55b26b5390d"} Oct 10 13:19:46 crc kubenswrapper[4745]: I1010 13:19:46.744508 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:46 crc kubenswrapper[4745]: I1010 13:19:46.744598 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:46 crc kubenswrapper[4745]: E1010 13:19:46.746501 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:46 crc kubenswrapper[4745]: E1010 13:19:46.746636 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:47 crc kubenswrapper[4745]: I1010 13:19:47.743953 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:47 crc kubenswrapper[4745]: I1010 13:19:47.743990 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:47 crc kubenswrapper[4745]: E1010 13:19:47.744165 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:47 crc kubenswrapper[4745]: E1010 13:19:47.744342 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:48 crc kubenswrapper[4745]: I1010 13:19:48.744906 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:48 crc kubenswrapper[4745]: I1010 13:19:48.744915 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:48 crc kubenswrapper[4745]: E1010 13:19:48.745133 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:48 crc kubenswrapper[4745]: E1010 13:19:48.745264 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:49 crc kubenswrapper[4745]: I1010 13:19:49.744085 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:49 crc kubenswrapper[4745]: I1010 13:19:49.744192 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:49 crc kubenswrapper[4745]: E1010 13:19:49.744243 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:49 crc kubenswrapper[4745]: E1010 13:19:49.744316 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:50 crc kubenswrapper[4745]: I1010 13:19:50.745126 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:50 crc kubenswrapper[4745]: I1010 13:19:50.745224 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:50 crc kubenswrapper[4745]: E1010 13:19:50.745562 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:50 crc kubenswrapper[4745]: E1010 13:19:50.745691 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:51 crc kubenswrapper[4745]: I1010 13:19:51.744800 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:51 crc kubenswrapper[4745]: I1010 13:19:51.744884 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:51 crc kubenswrapper[4745]: E1010 13:19:51.744921 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:51 crc kubenswrapper[4745]: E1010 13:19:51.745061 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:52 crc kubenswrapper[4745]: I1010 13:19:52.744803 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:52 crc kubenswrapper[4745]: I1010 13:19:52.744835 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:52 crc kubenswrapper[4745]: E1010 13:19:52.745004 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:52 crc kubenswrapper[4745]: E1010 13:19:52.745162 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:53 crc kubenswrapper[4745]: I1010 13:19:53.745100 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:53 crc kubenswrapper[4745]: I1010 13:19:53.745228 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:53 crc kubenswrapper[4745]: E1010 13:19:53.745302 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:53 crc kubenswrapper[4745]: E1010 13:19:53.745454 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:54 crc kubenswrapper[4745]: I1010 13:19:54.744495 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:54 crc kubenswrapper[4745]: I1010 13:19:54.744674 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:54 crc kubenswrapper[4745]: E1010 13:19:54.744892 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:54 crc kubenswrapper[4745]: E1010 13:19:54.745023 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:55 crc kubenswrapper[4745]: I1010 13:19:55.744922 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:55 crc kubenswrapper[4745]: I1010 13:19:55.745016 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:55 crc kubenswrapper[4745]: E1010 13:19:55.745697 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:55 crc kubenswrapper[4745]: E1010 13:19:55.745937 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:56 crc kubenswrapper[4745]: I1010 13:19:56.744293 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:56 crc kubenswrapper[4745]: I1010 13:19:56.744358 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:56 crc kubenswrapper[4745]: E1010 13:19:56.745337 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:56 crc kubenswrapper[4745]: E1010 13:19:56.745411 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:57 crc kubenswrapper[4745]: I1010 13:19:57.744260 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:57 crc kubenswrapper[4745]: I1010 13:19:57.744326 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:57 crc kubenswrapper[4745]: E1010 13:19:57.744391 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:57 crc kubenswrapper[4745]: E1010 13:19:57.744461 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:58 crc kubenswrapper[4745]: I1010 13:19:58.744138 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:19:58 crc kubenswrapper[4745]: I1010 13:19:58.744223 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:19:58 crc kubenswrapper[4745]: E1010 13:19:58.744301 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:19:58 crc kubenswrapper[4745]: E1010 13:19:58.744344 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:19:59 crc kubenswrapper[4745]: I1010 13:19:59.744302 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:19:59 crc kubenswrapper[4745]: I1010 13:19:59.744347 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:19:59 crc kubenswrapper[4745]: E1010 13:19:59.744511 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:19:59 crc kubenswrapper[4745]: E1010 13:19:59.745194 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:19:59 crc kubenswrapper[4745]: I1010 13:19:59.745962 4745 scope.go:117] "RemoveContainer" containerID="b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8" Oct 10 13:20:00 crc kubenswrapper[4745]: I1010 13:20:00.363818 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovnkube-controller/3.log" Oct 10 13:20:00 crc kubenswrapper[4745]: I1010 13:20:00.365857 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerStarted","Data":"668fb16e5543f24cd2b93f606627d511560b74d218f48fabe602ac3781ea471d"} Oct 10 13:20:00 crc kubenswrapper[4745]: I1010 13:20:00.366578 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:20:00 crc kubenswrapper[4745]: I1010 13:20:00.399247 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m78fg" podStartSLOduration=93.399220409 podStartE2EDuration="1m33.399220409s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:19:46.334183878 +0000 UTC m=+100.231840651" watchObservedRunningTime="2025-10-10 13:20:00.399220409 +0000 UTC m=+114.296877212" Oct 10 13:20:00 crc kubenswrapper[4745]: I1010 13:20:00.400445 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podStartSLOduration=93.400428513 podStartE2EDuration="1m33.400428513s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:00.399099955 +0000 UTC m=+114.296756728" watchObservedRunningTime="2025-10-10 13:20:00.400428513 +0000 UTC m=+114.298085316" Oct 10 13:20:00 crc kubenswrapper[4745]: I1010 13:20:00.659639 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xgh8p"] Oct 10 13:20:00 crc kubenswrapper[4745]: I1010 13:20:00.659817 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:20:00 crc kubenswrapper[4745]: E1010 13:20:00.659967 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:20:00 crc kubenswrapper[4745]: I1010 13:20:00.744261 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:20:00 crc kubenswrapper[4745]: E1010 13:20:00.744405 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:20:00 crc kubenswrapper[4745]: I1010 13:20:00.744512 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:20:00 crc kubenswrapper[4745]: E1010 13:20:00.744720 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:20:01 crc kubenswrapper[4745]: I1010 13:20:01.744618 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:20:01 crc kubenswrapper[4745]: E1010 13:20:01.744785 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 13:20:02 crc kubenswrapper[4745]: I1010 13:20:02.744589 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:20:02 crc kubenswrapper[4745]: E1010 13:20:02.744754 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xgh8p" podUID="82c4e63a-009f-4720-87b5-640a8ec65117" Oct 10 13:20:02 crc kubenswrapper[4745]: I1010 13:20:02.744827 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:20:02 crc kubenswrapper[4745]: I1010 13:20:02.744889 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:20:02 crc kubenswrapper[4745]: E1010 13:20:02.744977 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 13:20:02 crc kubenswrapper[4745]: E1010 13:20:02.745056 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.029214 4745 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.029427 4745 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.078647 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9q9xb"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.079272 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.079642 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bgmxs"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.079981 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.105477 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.105872 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.113391 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.114595 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.116107 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.117657 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-s8wbg"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.121640 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-72w5j"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.118207 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.118244 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.119815 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.119901 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.123353 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.126709 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.127136 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.128040 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.131367 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.129124 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.129341 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.129398 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.129673 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.130469 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.130632 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.132195 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.132218 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7v4cv"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.130687 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.130764 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.130815 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.132802 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dtjt9"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.130991 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.132904 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-7v4cv" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.131247 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.131374 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.131532 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.133700 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fqb4s"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.133820 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dtjt9" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.134112 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.134361 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fqb4s" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.134898 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kcwwb"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.135236 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.135763 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-98dn4"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.135984 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kcwwb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.136202 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.136650 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-7jjxn"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.136752 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.137214 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qss59"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.137274 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-7jjxn" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.137620 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.138145 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.138302 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.138465 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.138834 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.138882 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.139111 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.139248 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qs225"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.139265 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.139896 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.140504 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-76fgg"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.140884 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.141157 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-76fgg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.141201 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dfjd5"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.141573 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dfjd5" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.144772 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-5ftts"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.145153 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6gqzf"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.145464 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hqxgq"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.145958 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hqxgq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.146656 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-npfv4"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.158186 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6gqzf" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.159626 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.169468 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.169973 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.170144 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.171397 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.171408 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.171464 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.171518 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.171780 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.171877 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.172042 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.172079 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.172173 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.172367 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.172407 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.172588 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.172914 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.173033 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.173160 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.173191 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.173261 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.173364 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.173461 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.173566 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.173670 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.173792 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.174116 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.174224 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.174382 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-npfv4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.175740 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ncgfh"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.176469 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-xq5mm"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.176847 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.177307 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ncgfh" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.177337 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bgmxs"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.180081 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.180832 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.180982 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-kth6t"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.181169 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.181290 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.181757 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.181980 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.182107 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.182259 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.182394 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-kth6t" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.182424 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.183127 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jr8mx"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.183368 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.183563 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jr8mx" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.183864 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.183885 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.183978 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.184076 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.184122 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.184139 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.184001 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.184385 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.184310 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.184304 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.184352 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.184987 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8m4nd"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.185450 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.185658 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.185778 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.186097 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lwl2w"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.186510 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.186752 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lwl2w" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.186874 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.186882 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.186914 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.187043 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.187245 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.187667 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.187840 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.189534 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.192958 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dc27t"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.194351 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vrllq"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.195451 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vrllq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.195696 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.196413 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.196442 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.196694 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.196794 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.196927 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.196987 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.197006 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.197192 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.197395 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.197560 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.197599 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.197199 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.198304 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.198850 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.213521 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpmsh\" (UniqueName: \"kubernetes.io/projected/420a8490-c5b6-4d9b-a716-efc526b5f4a5-kube-api-access-jpmsh\") pod \"console-operator-58897d9998-7v4cv\" (UID: \"420a8490-c5b6-4d9b-a716-efc526b5f4a5\") " pod="openshift-console-operator/console-operator-58897d9998-7v4cv" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.213561 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c157d5c1-d9e5-4b0d-94b7-432bb85ffa10-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qs225\" (UID: \"c157d5c1-d9e5-4b0d-94b7-432bb85ffa10\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.213587 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfb47\" (UniqueName: \"kubernetes.io/projected/39f07a80-8096-439f-a3b5-b4dca2842b90-kube-api-access-qfb47\") pod \"openshift-config-operator-7777fb866f-76fgg\" (UID: \"39f07a80-8096-439f-a3b5-b4dca2842b90\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76fgg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.213614 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6s46\" (UniqueName: \"kubernetes.io/projected/8f35439c-e788-4e19-8054-6a365d28a76d-kube-api-access-z6s46\") pod \"cluster-samples-operator-665b6dd947-kcwwb\" (UID: \"8f35439c-e788-4e19-8054-6a365d28a76d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kcwwb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.213638 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-image-import-ca\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.213655 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-serving-cert\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.213678 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9-stats-auth\") pod \"router-default-5444994796-5ftts\" (UID: \"f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9\") " pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.213698 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hskss\" (UniqueName: \"kubernetes.io/projected/60290e46-da36-4e07-a1db-edc753eb0788-kube-api-access-hskss\") pod \"machine-approver-56656f9798-pg69q\" (UID: \"60290e46-da36-4e07-a1db-edc753eb0788\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.213723 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqbfm\" (UniqueName: \"kubernetes.io/projected/0088430d-3a96-46ea-a149-7c59eb24d706-kube-api-access-gqbfm\") pod \"controller-manager-879f6c89f-9q9xb\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.213761 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.213798 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9lsm\" (UniqueName: \"kubernetes.io/projected/839da6a0-320a-45d1-8a6c-be916ee27421-kube-api-access-t9lsm\") pod \"control-plane-machine-set-operator-78cbb6b69f-fqb4s\" (UID: \"839da6a0-320a-45d1-8a6c-be916ee27421\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fqb4s" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.213818 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f354b96e-de50-4ad1-8a82-60d570b1c6b5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.213833 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f354b96e-de50-4ad1-8a82-60d570b1c6b5-audit-dir\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.213942 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-encryption-config\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.213961 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-client-ca\") pod \"controller-manager-879f6c89f-9q9xb\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.213991 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-audit-policies\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214012 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9365b5c-a2b4-464c-aef6-afd71c5d3427-config\") pod \"route-controller-manager-6576b87f9c-nwvhl\" (UID: \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214047 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/25f468fc-1ca1-461d-a077-2150d42bcafe-proxy-tls\") pod \"machine-config-operator-74547568cd-rvnrp\" (UID: \"25f468fc-1ca1-461d-a077-2150d42bcafe\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214065 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f354b96e-de50-4ad1-8a82-60d570b1c6b5-serving-cert\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214083 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p524m\" (UniqueName: \"kubernetes.io/projected/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-kube-api-access-p524m\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214103 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/25f468fc-1ca1-461d-a077-2150d42bcafe-images\") pod \"machine-config-operator-74547568cd-rvnrp\" (UID: \"25f468fc-1ca1-461d-a077-2150d42bcafe\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214119 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfxkz\" (UniqueName: \"kubernetes.io/projected/f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75-kube-api-access-zfxkz\") pod \"cluster-image-registry-operator-dc59b4c8b-fk8ff\" (UID: \"f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214136 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f354b96e-de50-4ad1-8a82-60d570b1c6b5-etcd-client\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214153 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/800c6118-3c93-4dd4-b7e4-17f23b82de1d-config\") pod \"machine-api-operator-5694c8668f-s8wbg\" (UID: \"800c6118-3c93-4dd4-b7e4-17f23b82de1d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214171 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-config\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214187 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-audit\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214207 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptc8p\" (UniqueName: \"kubernetes.io/projected/247274a0-c2f0-4150-ad4f-cf39c2003a60-kube-api-access-ptc8p\") pod \"downloads-7954f5f757-7jjxn\" (UID: \"247274a0-c2f0-4150-ad4f-cf39c2003a60\") " pod="openshift-console/downloads-7954f5f757-7jjxn" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214226 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214246 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fk8ff\" (UID: \"f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214334 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9q9xb\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214350 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/25f468fc-1ca1-461d-a077-2150d42bcafe-auth-proxy-config\") pod \"machine-config-operator-74547568cd-rvnrp\" (UID: \"25f468fc-1ca1-461d-a077-2150d42bcafe\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214371 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/839da6a0-320a-45d1-8a6c-be916ee27421-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fqb4s\" (UID: \"839da6a0-320a-45d1-8a6c-be916ee27421\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fqb4s" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214407 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzrc9\" (UniqueName: \"kubernetes.io/projected/c157d5c1-d9e5-4b0d-94b7-432bb85ffa10-kube-api-access-rzrc9\") pod \"ingress-operator-5b745b69d9-qs225\" (UID: \"c157d5c1-d9e5-4b0d-94b7-432bb85ffa10\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214565 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214597 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f35439c-e788-4e19-8054-6a365d28a76d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-kcwwb\" (UID: \"8f35439c-e788-4e19-8054-6a365d28a76d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kcwwb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214722 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88252921-5e6b-4e51-95e8-59dd9dd891b6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bgmxs\" (UID: \"88252921-5e6b-4e51-95e8-59dd9dd891b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214794 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8nrx\" (UniqueName: \"kubernetes.io/projected/0b152ecd-75c1-4677-8f8a-c27ff65d76d9-kube-api-access-m8nrx\") pod \"migrator-59844c95c7-hqxgq\" (UID: \"0b152ecd-75c1-4677-8f8a-c27ff65d76d9\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hqxgq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214900 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d159f38f-c259-44cf-a5f9-8c4911d7864c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dtjt9\" (UID: \"d159f38f-c259-44cf-a5f9-8c4911d7864c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dtjt9" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214919 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c157d5c1-d9e5-4b0d-94b7-432bb85ffa10-metrics-tls\") pod \"ingress-operator-5b745b69d9-qs225\" (UID: \"c157d5c1-d9e5-4b0d-94b7-432bb85ffa10\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.214958 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-etcd-serving-ca\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.215083 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-config\") pod \"controller-manager-879f6c89f-9q9xb\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.215119 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0088430d-3a96-46ea-a149-7c59eb24d706-serving-cert\") pod \"controller-manager-879f6c89f-9q9xb\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.215210 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkdqw\" (UniqueName: \"kubernetes.io/projected/f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9-kube-api-access-pkdqw\") pod \"router-default-5444994796-5ftts\" (UID: \"f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9\") " pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.215228 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/800c6118-3c93-4dd4-b7e4-17f23b82de1d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-s8wbg\" (UID: \"800c6118-3c93-4dd4-b7e4-17f23b82de1d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.215335 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lfpb\" (UniqueName: \"kubernetes.io/projected/7ddc713d-4db8-4e77-acd2-0936a8e7bcb1-kube-api-access-8lfpb\") pod \"openshift-controller-manager-operator-756b6f6bc6-dfjd5\" (UID: \"7ddc713d-4db8-4e77-acd2-0936a8e7bcb1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dfjd5" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.218344 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.218761 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.220401 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfw7c\" (UniqueName: \"kubernetes.io/projected/25f468fc-1ca1-461d-a077-2150d42bcafe-kube-api-access-mfw7c\") pod \"machine-config-operator-74547568cd-rvnrp\" (UID: \"25f468fc-1ca1-461d-a077-2150d42bcafe\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.220509 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbqrq\" (UniqueName: \"kubernetes.io/projected/800c6118-3c93-4dd4-b7e4-17f23b82de1d-kube-api-access-lbqrq\") pod \"machine-api-operator-5694c8668f-s8wbg\" (UID: \"800c6118-3c93-4dd4-b7e4-17f23b82de1d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.220563 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67d48\" (UniqueName: \"kubernetes.io/projected/88252921-5e6b-4e51-95e8-59dd9dd891b6-kube-api-access-67d48\") pod \"authentication-operator-69f744f599-bgmxs\" (UID: \"88252921-5e6b-4e51-95e8-59dd9dd891b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.220647 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9-service-ca-bundle\") pod \"router-default-5444994796-5ftts\" (UID: \"f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9\") " pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.220677 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0c14e43-fd09-45b2-9820-abb38a0e7761-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6gqzf\" (UID: \"e0c14e43-fd09-45b2-9820-abb38a0e7761\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6gqzf" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.220697 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-audit-dir\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.220736 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.220808 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9365b5c-a2b4-464c-aef6-afd71c5d3427-serving-cert\") pod \"route-controller-manager-6576b87f9c-nwvhl\" (UID: \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.220867 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0c14e43-fd09-45b2-9820-abb38a0e7761-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6gqzf\" (UID: \"e0c14e43-fd09-45b2-9820-abb38a0e7761\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6gqzf" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.220888 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c157d5c1-d9e5-4b0d-94b7-432bb85ffa10-trusted-ca\") pod \"ingress-operator-5b745b69d9-qs225\" (UID: \"c157d5c1-d9e5-4b0d-94b7-432bb85ffa10\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.220906 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9365b5c-a2b4-464c-aef6-afd71c5d3427-client-ca\") pod \"route-controller-manager-6576b87f9c-nwvhl\" (UID: \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.220937 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/60290e46-da36-4e07-a1db-edc753eb0788-auth-proxy-config\") pod \"machine-approver-56656f9798-pg69q\" (UID: \"60290e46-da36-4e07-a1db-edc753eb0788\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.220968 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ddc713d-4db8-4e77-acd2-0936a8e7bcb1-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-dfjd5\" (UID: \"7ddc713d-4db8-4e77-acd2-0936a8e7bcb1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dfjd5" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.221017 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.221143 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d159f38f-c259-44cf-a5f9-8c4911d7864c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dtjt9\" (UID: \"d159f38f-c259-44cf-a5f9-8c4911d7864c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dtjt9" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.221179 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/420a8490-c5b6-4d9b-a716-efc526b5f4a5-trusted-ca\") pod \"console-operator-58897d9998-7v4cv\" (UID: \"420a8490-c5b6-4d9b-a716-efc526b5f4a5\") " pod="openshift-console-operator/console-operator-58897d9998-7v4cv" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.221385 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.221474 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60290e46-da36-4e07-a1db-edc753eb0788-config\") pod \"machine-approver-56656f9798-pg69q\" (UID: \"60290e46-da36-4e07-a1db-edc753eb0788\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.221529 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/420a8490-c5b6-4d9b-a716-efc526b5f4a5-config\") pod \"console-operator-58897d9998-7v4cv\" (UID: \"420a8490-c5b6-4d9b-a716-efc526b5f4a5\") " pod="openshift-console-operator/console-operator-58897d9998-7v4cv" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.221696 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88252921-5e6b-4e51-95e8-59dd9dd891b6-serving-cert\") pod \"authentication-operator-69f744f599-bgmxs\" (UID: \"88252921-5e6b-4e51-95e8-59dd9dd891b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.221967 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f354b96e-de50-4ad1-8a82-60d570b1c6b5-encryption-config\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.222015 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-etcd-client\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.222036 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/420a8490-c5b6-4d9b-a716-efc526b5f4a5-serving-cert\") pod \"console-operator-58897d9998-7v4cv\" (UID: \"420a8490-c5b6-4d9b-a716-efc526b5f4a5\") " pod="openshift-console-operator/console-operator-58897d9998-7v4cv" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.222149 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.222209 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.222261 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk4gk\" (UniqueName: \"kubernetes.io/projected/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-kube-api-access-tk4gk\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.222333 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.222436 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/39f07a80-8096-439f-a3b5-b4dca2842b90-available-featuregates\") pod \"openshift-config-operator-7777fb866f-76fgg\" (UID: \"39f07a80-8096-439f-a3b5-b4dca2842b90\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76fgg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.222765 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88252921-5e6b-4e51-95e8-59dd9dd891b6-config\") pod \"authentication-operator-69f744f599-bgmxs\" (UID: \"88252921-5e6b-4e51-95e8-59dd9dd891b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.222844 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/60290e46-da36-4e07-a1db-edc753eb0788-machine-approver-tls\") pod \"machine-approver-56656f9798-pg69q\" (UID: \"60290e46-da36-4e07-a1db-edc753eb0788\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.240883 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.246022 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.246912 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.247102 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.247226 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j6gkw"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.247268 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9-metrics-certs\") pod \"router-default-5444994796-5ftts\" (UID: \"f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9\") " pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.247369 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqrdv\" (UniqueName: \"kubernetes.io/projected/e0c14e43-fd09-45b2-9820-abb38a0e7761-kube-api-access-dqrdv\") pod \"kube-storage-version-migrator-operator-b67b599dd-6gqzf\" (UID: \"e0c14e43-fd09-45b2-9820-abb38a0e7761\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6gqzf" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.247406 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f354b96e-de50-4ad1-8a82-60d570b1c6b5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.247453 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-audit-dir\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.247478 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9-default-certificate\") pod \"router-default-5444994796-5ftts\" (UID: \"f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9\") " pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.247531 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fk8ff\" (UID: \"f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.247695 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fk8ff\" (UID: \"f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.247831 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.247871 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/800c6118-3c93-4dd4-b7e4-17f23b82de1d-images\") pod \"machine-api-operator-5694c8668f-s8wbg\" (UID: \"800c6118-3c93-4dd4-b7e4-17f23b82de1d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.247901 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.247938 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ntfd\" (UniqueName: \"kubernetes.io/projected/d159f38f-c259-44cf-a5f9-8c4911d7864c-kube-api-access-4ntfd\") pod \"openshift-apiserver-operator-796bbdcf4f-dtjt9\" (UID: \"d159f38f-c259-44cf-a5f9-8c4911d7864c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dtjt9" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.247961 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ddc713d-4db8-4e77-acd2-0936a8e7bcb1-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-dfjd5\" (UID: \"7ddc713d-4db8-4e77-acd2-0936a8e7bcb1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dfjd5" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.247987 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.247995 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f354b96e-de50-4ad1-8a82-60d570b1c6b5-audit-policies\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.248016 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.248033 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39f07a80-8096-439f-a3b5-b4dca2842b90-serving-cert\") pod \"openshift-config-operator-7777fb866f-76fgg\" (UID: \"39f07a80-8096-439f-a3b5-b4dca2842b90\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76fgg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.248054 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zgdj\" (UniqueName: \"kubernetes.io/projected/f9365b5c-a2b4-464c-aef6-afd71c5d3427-kube-api-access-5zgdj\") pod \"route-controller-manager-6576b87f9c-nwvhl\" (UID: \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.248083 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.248093 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.248345 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.248518 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j6gkw" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.249112 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.249720 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88252921-5e6b-4e51-95e8-59dd9dd891b6-service-ca-bundle\") pod \"authentication-operator-69f744f599-bgmxs\" (UID: \"88252921-5e6b-4e51-95e8-59dd9dd891b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.251192 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.251590 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-kgpj4"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.252171 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-kgpj4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.253877 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-72w5j"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.253916 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9q9xb"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.254819 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjnxd\" (UniqueName: \"kubernetes.io/projected/f354b96e-de50-4ad1-8a82-60d570b1c6b5-kube-api-access-jjnxd\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.254863 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-trusted-ca-bundle\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.254886 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-node-pullsecrets\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.258224 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.263157 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.264197 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.264448 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kcwwb"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.266724 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-s8wbg"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.266778 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.267002 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.267613 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.269839 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fqb4s"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.271981 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qss59"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.273688 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.273884 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-76fgg"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.274891 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dfjd5"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.276824 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-rwdj4"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.278096 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-kth6t"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.278177 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.279367 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-7jjxn"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.281911 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.282548 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jr8mx"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.284257 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qs225"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.286353 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dc27t"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.287765 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hqxgq"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.289139 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7v4cv"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.290702 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-npfv4"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.291999 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ncgfh"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.294607 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6gqzf"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.301428 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.303027 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.304884 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dtjt9"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.305757 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-98dn4"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.306643 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lwl2w"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.307652 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.308637 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vrllq"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.309938 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.310917 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-xq5mm"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.311937 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8m4nd"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.315783 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.318020 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.319850 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-pdt5f"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.320721 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-pdt5f" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.321079 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-gffs7"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.321924 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-gffs7" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.322078 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.322790 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-kgpj4"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.323620 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-gffs7"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.326304 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.333405 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j6gkw"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.335048 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-rwdj4"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.335769 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-zdbqc"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.336497 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-zdbqc" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.336531 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-zdbqc"] Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.341286 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355472 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9365b5c-a2b4-464c-aef6-afd71c5d3427-client-ca\") pod \"route-controller-manager-6576b87f9c-nwvhl\" (UID: \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355498 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9365b5c-a2b4-464c-aef6-afd71c5d3427-serving-cert\") pod \"route-controller-manager-6576b87f9c-nwvhl\" (UID: \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355516 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0c14e43-fd09-45b2-9820-abb38a0e7761-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6gqzf\" (UID: \"e0c14e43-fd09-45b2-9820-abb38a0e7761\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6gqzf" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355533 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c157d5c1-d9e5-4b0d-94b7-432bb85ffa10-trusted-ca\") pod \"ingress-operator-5b745b69d9-qs225\" (UID: \"c157d5c1-d9e5-4b0d-94b7-432bb85ffa10\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355548 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/60290e46-da36-4e07-a1db-edc753eb0788-auth-proxy-config\") pod \"machine-approver-56656f9798-pg69q\" (UID: \"60290e46-da36-4e07-a1db-edc753eb0788\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355564 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ddc713d-4db8-4e77-acd2-0936a8e7bcb1-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-dfjd5\" (UID: \"7ddc713d-4db8-4e77-acd2-0936a8e7bcb1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dfjd5" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355586 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/420a8490-c5b6-4d9b-a716-efc526b5f4a5-trusted-ca\") pod \"console-operator-58897d9998-7v4cv\" (UID: \"420a8490-c5b6-4d9b-a716-efc526b5f4a5\") " pod="openshift-console-operator/console-operator-58897d9998-7v4cv" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355604 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355620 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d159f38f-c259-44cf-a5f9-8c4911d7864c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dtjt9\" (UID: \"d159f38f-c259-44cf-a5f9-8c4911d7864c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dtjt9" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355637 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60290e46-da36-4e07-a1db-edc753eb0788-config\") pod \"machine-approver-56656f9798-pg69q\" (UID: \"60290e46-da36-4e07-a1db-edc753eb0788\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355650 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/420a8490-c5b6-4d9b-a716-efc526b5f4a5-config\") pod \"console-operator-58897d9998-7v4cv\" (UID: \"420a8490-c5b6-4d9b-a716-efc526b5f4a5\") " pod="openshift-console-operator/console-operator-58897d9998-7v4cv" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355665 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88252921-5e6b-4e51-95e8-59dd9dd891b6-serving-cert\") pod \"authentication-operator-69f744f599-bgmxs\" (UID: \"88252921-5e6b-4e51-95e8-59dd9dd891b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355688 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f354b96e-de50-4ad1-8a82-60d570b1c6b5-encryption-config\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355709 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-etcd-client\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355737 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/420a8490-c5b6-4d9b-a716-efc526b5f4a5-serving-cert\") pod \"console-operator-58897d9998-7v4cv\" (UID: \"420a8490-c5b6-4d9b-a716-efc526b5f4a5\") " pod="openshift-console-operator/console-operator-58897d9998-7v4cv" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355753 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355767 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355786 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk4gk\" (UniqueName: \"kubernetes.io/projected/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-kube-api-access-tk4gk\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355803 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355818 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/39f07a80-8096-439f-a3b5-b4dca2842b90-available-featuregates\") pod \"openshift-config-operator-7777fb866f-76fgg\" (UID: \"39f07a80-8096-439f-a3b5-b4dca2842b90\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76fgg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355840 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88252921-5e6b-4e51-95e8-59dd9dd891b6-config\") pod \"authentication-operator-69f744f599-bgmxs\" (UID: \"88252921-5e6b-4e51-95e8-59dd9dd891b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355858 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f354b96e-de50-4ad1-8a82-60d570b1c6b5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355873 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/60290e46-da36-4e07-a1db-edc753eb0788-machine-approver-tls\") pod \"machine-approver-56656f9798-pg69q\" (UID: \"60290e46-da36-4e07-a1db-edc753eb0788\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355888 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9-metrics-certs\") pod \"router-default-5444994796-5ftts\" (UID: \"f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9\") " pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355904 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqrdv\" (UniqueName: \"kubernetes.io/projected/e0c14e43-fd09-45b2-9820-abb38a0e7761-kube-api-access-dqrdv\") pod \"kube-storage-version-migrator-operator-b67b599dd-6gqzf\" (UID: \"e0c14e43-fd09-45b2-9820-abb38a0e7761\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6gqzf" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355926 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-audit-dir\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355941 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9-default-certificate\") pod \"router-default-5444994796-5ftts\" (UID: \"f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9\") " pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355964 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/800c6118-3c93-4dd4-b7e4-17f23b82de1d-images\") pod \"machine-api-operator-5694c8668f-s8wbg\" (UID: \"800c6118-3c93-4dd4-b7e4-17f23b82de1d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.355988 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fk8ff\" (UID: \"f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356008 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fk8ff\" (UID: \"f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356028 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ntfd\" (UniqueName: \"kubernetes.io/projected/d159f38f-c259-44cf-a5f9-8c4911d7864c-kube-api-access-4ntfd\") pod \"openshift-apiserver-operator-796bbdcf4f-dtjt9\" (UID: \"d159f38f-c259-44cf-a5f9-8c4911d7864c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dtjt9" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356102 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356141 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ddc713d-4db8-4e77-acd2-0936a8e7bcb1-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-dfjd5\" (UID: \"7ddc713d-4db8-4e77-acd2-0936a8e7bcb1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dfjd5" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356159 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f354b96e-de50-4ad1-8a82-60d570b1c6b5-audit-policies\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356173 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356189 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39f07a80-8096-439f-a3b5-b4dca2842b90-serving-cert\") pod \"openshift-config-operator-7777fb866f-76fgg\" (UID: \"39f07a80-8096-439f-a3b5-b4dca2842b90\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76fgg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356204 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zgdj\" (UniqueName: \"kubernetes.io/projected/f9365b5c-a2b4-464c-aef6-afd71c5d3427-kube-api-access-5zgdj\") pod \"route-controller-manager-6576b87f9c-nwvhl\" (UID: \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356221 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356236 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88252921-5e6b-4e51-95e8-59dd9dd891b6-service-ca-bundle\") pod \"authentication-operator-69f744f599-bgmxs\" (UID: \"88252921-5e6b-4e51-95e8-59dd9dd891b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356251 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjnxd\" (UniqueName: \"kubernetes.io/projected/f354b96e-de50-4ad1-8a82-60d570b1c6b5-kube-api-access-jjnxd\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356269 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-trusted-ca-bundle\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356285 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-node-pullsecrets\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356303 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpmsh\" (UniqueName: \"kubernetes.io/projected/420a8490-c5b6-4d9b-a716-efc526b5f4a5-kube-api-access-jpmsh\") pod \"console-operator-58897d9998-7v4cv\" (UID: \"420a8490-c5b6-4d9b-a716-efc526b5f4a5\") " pod="openshift-console-operator/console-operator-58897d9998-7v4cv" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356317 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c157d5c1-d9e5-4b0d-94b7-432bb85ffa10-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qs225\" (UID: \"c157d5c1-d9e5-4b0d-94b7-432bb85ffa10\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356370 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9365b5c-a2b4-464c-aef6-afd71c5d3427-client-ca\") pod \"route-controller-manager-6576b87f9c-nwvhl\" (UID: \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356409 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfb47\" (UniqueName: \"kubernetes.io/projected/39f07a80-8096-439f-a3b5-b4dca2842b90-kube-api-access-qfb47\") pod \"openshift-config-operator-7777fb866f-76fgg\" (UID: \"39f07a80-8096-439f-a3b5-b4dca2842b90\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76fgg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356975 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6s46\" (UniqueName: \"kubernetes.io/projected/8f35439c-e788-4e19-8054-6a365d28a76d-kube-api-access-z6s46\") pod \"cluster-samples-operator-665b6dd947-kcwwb\" (UID: \"8f35439c-e788-4e19-8054-6a365d28a76d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kcwwb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356998 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-image-import-ca\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357017 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-serving-cert\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357035 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9-stats-auth\") pod \"router-default-5444994796-5ftts\" (UID: \"f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9\") " pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357054 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hskss\" (UniqueName: \"kubernetes.io/projected/60290e46-da36-4e07-a1db-edc753eb0788-kube-api-access-hskss\") pod \"machine-approver-56656f9798-pg69q\" (UID: \"60290e46-da36-4e07-a1db-edc753eb0788\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357072 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqbfm\" (UniqueName: \"kubernetes.io/projected/0088430d-3a96-46ea-a149-7c59eb24d706-kube-api-access-gqbfm\") pod \"controller-manager-879f6c89f-9q9xb\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357091 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357121 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9lsm\" (UniqueName: \"kubernetes.io/projected/839da6a0-320a-45d1-8a6c-be916ee27421-kube-api-access-t9lsm\") pod \"control-plane-machine-set-operator-78cbb6b69f-fqb4s\" (UID: \"839da6a0-320a-45d1-8a6c-be916ee27421\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fqb4s" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357140 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f354b96e-de50-4ad1-8a82-60d570b1c6b5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357154 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f354b96e-de50-4ad1-8a82-60d570b1c6b5-audit-dir\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357170 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-encryption-config\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357185 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-client-ca\") pod \"controller-manager-879f6c89f-9q9xb\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357205 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/25f468fc-1ca1-461d-a077-2150d42bcafe-proxy-tls\") pod \"machine-config-operator-74547568cd-rvnrp\" (UID: \"25f468fc-1ca1-461d-a077-2150d42bcafe\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357221 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-audit-policies\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357235 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9365b5c-a2b4-464c-aef6-afd71c5d3427-config\") pod \"route-controller-manager-6576b87f9c-nwvhl\" (UID: \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357251 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f354b96e-de50-4ad1-8a82-60d570b1c6b5-serving-cert\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357268 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p524m\" (UniqueName: \"kubernetes.io/projected/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-kube-api-access-p524m\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357290 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/25f468fc-1ca1-461d-a077-2150d42bcafe-images\") pod \"machine-config-operator-74547568cd-rvnrp\" (UID: \"25f468fc-1ca1-461d-a077-2150d42bcafe\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357307 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfxkz\" (UniqueName: \"kubernetes.io/projected/f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75-kube-api-access-zfxkz\") pod \"cluster-image-registry-operator-dc59b4c8b-fk8ff\" (UID: \"f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357326 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f354b96e-de50-4ad1-8a82-60d570b1c6b5-etcd-client\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357342 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/800c6118-3c93-4dd4-b7e4-17f23b82de1d-config\") pod \"machine-api-operator-5694c8668f-s8wbg\" (UID: \"800c6118-3c93-4dd4-b7e4-17f23b82de1d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357358 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-config\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357374 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-audit\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357399 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptc8p\" (UniqueName: \"kubernetes.io/projected/247274a0-c2f0-4150-ad4f-cf39c2003a60-kube-api-access-ptc8p\") pod \"downloads-7954f5f757-7jjxn\" (UID: \"247274a0-c2f0-4150-ad4f-cf39c2003a60\") " pod="openshift-console/downloads-7954f5f757-7jjxn" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357420 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357440 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fk8ff\" (UID: \"f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357459 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/839da6a0-320a-45d1-8a6c-be916ee27421-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fqb4s\" (UID: \"839da6a0-320a-45d1-8a6c-be916ee27421\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fqb4s" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357476 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9q9xb\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357493 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/25f468fc-1ca1-461d-a077-2150d42bcafe-auth-proxy-config\") pod \"machine-config-operator-74547568cd-rvnrp\" (UID: \"25f468fc-1ca1-461d-a077-2150d42bcafe\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357511 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357530 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzrc9\" (UniqueName: \"kubernetes.io/projected/c157d5c1-d9e5-4b0d-94b7-432bb85ffa10-kube-api-access-rzrc9\") pod \"ingress-operator-5b745b69d9-qs225\" (UID: \"c157d5c1-d9e5-4b0d-94b7-432bb85ffa10\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357546 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88252921-5e6b-4e51-95e8-59dd9dd891b6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bgmxs\" (UID: \"88252921-5e6b-4e51-95e8-59dd9dd891b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357562 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f35439c-e788-4e19-8054-6a365d28a76d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-kcwwb\" (UID: \"8f35439c-e788-4e19-8054-6a365d28a76d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kcwwb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357583 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8nrx\" (UniqueName: \"kubernetes.io/projected/0b152ecd-75c1-4677-8f8a-c27ff65d76d9-kube-api-access-m8nrx\") pod \"migrator-59844c95c7-hqxgq\" (UID: \"0b152ecd-75c1-4677-8f8a-c27ff65d76d9\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hqxgq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357601 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d159f38f-c259-44cf-a5f9-8c4911d7864c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dtjt9\" (UID: \"d159f38f-c259-44cf-a5f9-8c4911d7864c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dtjt9" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357606 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/420a8490-c5b6-4d9b-a716-efc526b5f4a5-trusted-ca\") pod \"console-operator-58897d9998-7v4cv\" (UID: \"420a8490-c5b6-4d9b-a716-efc526b5f4a5\") " pod="openshift-console-operator/console-operator-58897d9998-7v4cv" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357618 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c157d5c1-d9e5-4b0d-94b7-432bb85ffa10-metrics-tls\") pod \"ingress-operator-5b745b69d9-qs225\" (UID: \"c157d5c1-d9e5-4b0d-94b7-432bb85ffa10\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357619 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/420a8490-c5b6-4d9b-a716-efc526b5f4a5-config\") pod \"console-operator-58897d9998-7v4cv\" (UID: \"420a8490-c5b6-4d9b-a716-efc526b5f4a5\") " pod="openshift-console-operator/console-operator-58897d9998-7v4cv" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357638 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/800c6118-3c93-4dd4-b7e4-17f23b82de1d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-s8wbg\" (UID: \"800c6118-3c93-4dd4-b7e4-17f23b82de1d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357665 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-etcd-serving-ca\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357687 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-config\") pod \"controller-manager-879f6c89f-9q9xb\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357705 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0088430d-3a96-46ea-a149-7c59eb24d706-serving-cert\") pod \"controller-manager-879f6c89f-9q9xb\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357723 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkdqw\" (UniqueName: \"kubernetes.io/projected/f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9-kube-api-access-pkdqw\") pod \"router-default-5444994796-5ftts\" (UID: \"f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9\") " pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357756 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbqrq\" (UniqueName: \"kubernetes.io/projected/800c6118-3c93-4dd4-b7e4-17f23b82de1d-kube-api-access-lbqrq\") pod \"machine-api-operator-5694c8668f-s8wbg\" (UID: \"800c6118-3c93-4dd4-b7e4-17f23b82de1d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357774 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lfpb\" (UniqueName: \"kubernetes.io/projected/7ddc713d-4db8-4e77-acd2-0936a8e7bcb1-kube-api-access-8lfpb\") pod \"openshift-controller-manager-operator-756b6f6bc6-dfjd5\" (UID: \"7ddc713d-4db8-4e77-acd2-0936a8e7bcb1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dfjd5" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357793 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfw7c\" (UniqueName: \"kubernetes.io/projected/25f468fc-1ca1-461d-a077-2150d42bcafe-kube-api-access-mfw7c\") pod \"machine-config-operator-74547568cd-rvnrp\" (UID: \"25f468fc-1ca1-461d-a077-2150d42bcafe\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357809 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357826 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67d48\" (UniqueName: \"kubernetes.io/projected/88252921-5e6b-4e51-95e8-59dd9dd891b6-kube-api-access-67d48\") pod \"authentication-operator-69f744f599-bgmxs\" (UID: \"88252921-5e6b-4e51-95e8-59dd9dd891b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357842 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9-service-ca-bundle\") pod \"router-default-5444994796-5ftts\" (UID: \"f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9\") " pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357858 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0c14e43-fd09-45b2-9820-abb38a0e7761-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6gqzf\" (UID: \"e0c14e43-fd09-45b2-9820-abb38a0e7761\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6gqzf" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357874 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-audit-dir\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.357937 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-audit-dir\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.358490 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d159f38f-c259-44cf-a5f9-8c4911d7864c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dtjt9\" (UID: \"d159f38f-c259-44cf-a5f9-8c4911d7864c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dtjt9" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.358930 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60290e46-da36-4e07-a1db-edc753eb0788-config\") pod \"machine-approver-56656f9798-pg69q\" (UID: \"60290e46-da36-4e07-a1db-edc753eb0788\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.359597 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-etcd-serving-ca\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.360756 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-config\") pod \"controller-manager-879f6c89f-9q9xb\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.361200 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f354b96e-de50-4ad1-8a82-60d570b1c6b5-encryption-config\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.361206 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39f07a80-8096-439f-a3b5-b4dca2842b90-serving-cert\") pod \"openshift-config-operator-7777fb866f-76fgg\" (UID: \"39f07a80-8096-439f-a3b5-b4dca2842b90\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76fgg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.361280 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.361386 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-audit-dir\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.361456 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/420a8490-c5b6-4d9b-a716-efc526b5f4a5-serving-cert\") pod \"console-operator-58897d9998-7v4cv\" (UID: \"420a8490-c5b6-4d9b-a716-efc526b5f4a5\") " pod="openshift-console-operator/console-operator-58897d9998-7v4cv" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.361752 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.361946 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-trusted-ca-bundle\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.362038 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-node-pullsecrets\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.362188 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/800c6118-3c93-4dd4-b7e4-17f23b82de1d-images\") pod \"machine-api-operator-5694c8668f-s8wbg\" (UID: \"800c6118-3c93-4dd4-b7e4-17f23b82de1d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.362294 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ddc713d-4db8-4e77-acd2-0936a8e7bcb1-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-dfjd5\" (UID: \"7ddc713d-4db8-4e77-acd2-0936a8e7bcb1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dfjd5" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356221 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/60290e46-da36-4e07-a1db-edc753eb0788-auth-proxy-config\") pod \"machine-approver-56656f9798-pg69q\" (UID: \"60290e46-da36-4e07-a1db-edc753eb0788\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.362603 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.362831 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.356697 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c157d5c1-d9e5-4b0d-94b7-432bb85ffa10-trusted-ca\") pod \"ingress-operator-5b745b69d9-qs225\" (UID: \"c157d5c1-d9e5-4b0d-94b7-432bb85ffa10\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.363837 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/60290e46-da36-4e07-a1db-edc753eb0788-machine-approver-tls\") pod \"machine-approver-56656f9798-pg69q\" (UID: \"60290e46-da36-4e07-a1db-edc753eb0788\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.364865 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88252921-5e6b-4e51-95e8-59dd9dd891b6-config\") pod \"authentication-operator-69f744f599-bgmxs\" (UID: \"88252921-5e6b-4e51-95e8-59dd9dd891b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.365062 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9365b5c-a2b4-464c-aef6-afd71c5d3427-serving-cert\") pod \"route-controller-manager-6576b87f9c-nwvhl\" (UID: \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.365377 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-config\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.365397 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/800c6118-3c93-4dd4-b7e4-17f23b82de1d-config\") pod \"machine-api-operator-5694c8668f-s8wbg\" (UID: \"800c6118-3c93-4dd4-b7e4-17f23b82de1d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.365599 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-audit\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.365700 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/800c6118-3c93-4dd4-b7e4-17f23b82de1d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-s8wbg\" (UID: \"800c6118-3c93-4dd4-b7e4-17f23b82de1d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.365838 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.366429 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88252921-5e6b-4e51-95e8-59dd9dd891b6-service-ca-bundle\") pod \"authentication-operator-69f744f599-bgmxs\" (UID: \"88252921-5e6b-4e51-95e8-59dd9dd891b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.366440 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/25f468fc-1ca1-461d-a077-2150d42bcafe-images\") pod \"machine-config-operator-74547568cd-rvnrp\" (UID: \"25f468fc-1ca1-461d-a077-2150d42bcafe\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.366489 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f354b96e-de50-4ad1-8a82-60d570b1c6b5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.366853 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-image-import-ca\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.367246 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88252921-5e6b-4e51-95e8-59dd9dd891b6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bgmxs\" (UID: \"88252921-5e6b-4e51-95e8-59dd9dd891b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.367334 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.367414 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fk8ff\" (UID: \"f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.367530 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9q9xb\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.367999 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ddc713d-4db8-4e77-acd2-0936a8e7bcb1-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-dfjd5\" (UID: \"7ddc713d-4db8-4e77-acd2-0936a8e7bcb1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dfjd5" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.368039 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-client-ca\") pod \"controller-manager-879f6c89f-9q9xb\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.368068 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/25f468fc-1ca1-461d-a077-2150d42bcafe-auth-proxy-config\") pod \"machine-config-operator-74547568cd-rvnrp\" (UID: \"25f468fc-1ca1-461d-a077-2150d42bcafe\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.368903 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0088430d-3a96-46ea-a149-7c59eb24d706-serving-cert\") pod \"controller-manager-879f6c89f-9q9xb\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.368905 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-serving-cert\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.369150 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.369410 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f354b96e-de50-4ad1-8a82-60d570b1c6b5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.369461 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88252921-5e6b-4e51-95e8-59dd9dd891b6-serving-cert\") pod \"authentication-operator-69f744f599-bgmxs\" (UID: \"88252921-5e6b-4e51-95e8-59dd9dd891b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.369577 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.369641 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f354b96e-de50-4ad1-8a82-60d570b1c6b5-audit-dir\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.369905 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.369983 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-audit-policies\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.370210 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/39f07a80-8096-439f-a3b5-b4dca2842b90-available-featuregates\") pod \"openshift-config-operator-7777fb866f-76fgg\" (UID: \"39f07a80-8096-439f-a3b5-b4dca2842b90\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76fgg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.370484 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f354b96e-de50-4ad1-8a82-60d570b1c6b5-audit-policies\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.370644 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9365b5c-a2b4-464c-aef6-afd71c5d3427-config\") pod \"route-controller-manager-6576b87f9c-nwvhl\" (UID: \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.370673 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.370813 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-encryption-config\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.370886 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-etcd-client\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.371238 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/25f468fc-1ca1-461d-a077-2150d42bcafe-proxy-tls\") pod \"machine-config-operator-74547568cd-rvnrp\" (UID: \"25f468fc-1ca1-461d-a077-2150d42bcafe\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.371348 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fk8ff\" (UID: \"f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.371416 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d159f38f-c259-44cf-a5f9-8c4911d7864c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dtjt9\" (UID: \"d159f38f-c259-44cf-a5f9-8c4911d7864c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dtjt9" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.371675 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f35439c-e788-4e19-8054-6a365d28a76d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-kcwwb\" (UID: \"8f35439c-e788-4e19-8054-6a365d28a76d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kcwwb" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.371800 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.372381 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.372666 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f354b96e-de50-4ad1-8a82-60d570b1c6b5-serving-cert\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.373568 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c157d5c1-d9e5-4b0d-94b7-432bb85ffa10-metrics-tls\") pod \"ingress-operator-5b745b69d9-qs225\" (UID: \"c157d5c1-d9e5-4b0d-94b7-432bb85ffa10\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.376426 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/839da6a0-320a-45d1-8a6c-be916ee27421-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fqb4s\" (UID: \"839da6a0-320a-45d1-8a6c-be916ee27421\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fqb4s" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.377626 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f354b96e-de50-4ad1-8a82-60d570b1c6b5-etcd-client\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.382340 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.401639 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.423029 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.429316 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0c14e43-fd09-45b2-9820-abb38a0e7761-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6gqzf\" (UID: \"e0c14e43-fd09-45b2-9820-abb38a0e7761\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6gqzf" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.442094 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.446200 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0c14e43-fd09-45b2-9820-abb38a0e7761-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6gqzf\" (UID: \"e0c14e43-fd09-45b2-9820-abb38a0e7761\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6gqzf" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.461494 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.481159 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.501389 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.509368 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9-stats-auth\") pod \"router-default-5444994796-5ftts\" (UID: \"f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9\") " pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.521195 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.530523 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9-metrics-certs\") pod \"router-default-5444994796-5ftts\" (UID: \"f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9\") " pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.541506 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.561603 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.575976 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9-default-certificate\") pod \"router-default-5444994796-5ftts\" (UID: \"f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9\") " pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.581640 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.586832 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9-service-ca-bundle\") pod \"router-default-5444994796-5ftts\" (UID: \"f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9\") " pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.601544 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.621852 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.661160 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.681343 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.701914 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.723894 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.741978 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.744305 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.761445 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.782049 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.801958 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.822140 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.842265 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.861112 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.881629 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.907340 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.921577 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.942182 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.961983 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 10 13:20:03 crc kubenswrapper[4745]: I1010 13:20:03.982102 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.001858 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.022369 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.042706 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.062332 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.082158 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.102560 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.122041 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.142263 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.161655 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.182309 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.200381 4745 request.go:700] Waited for 1.015544788s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-multus/secrets?fieldSelector=metadata.name%3Dmultus-admission-controller-secret&limit=500&resourceVersion=0 Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.201954 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.221630 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.241619 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.262157 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.280993 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.301488 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.321248 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.342007 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.361867 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.403175 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.421867 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.441541 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.461162 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.481257 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.501065 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.522012 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.542221 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.562087 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.582434 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.603026 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.621287 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.642542 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.671256 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.682181 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.701986 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.722492 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.741867 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.745007 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.745082 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.745011 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.762495 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.781653 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.802123 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.821883 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.842055 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.862790 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.882406 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.902148 4745 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.922387 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.943047 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.962276 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 10 13:20:04 crc kubenswrapper[4745]: I1010 13:20:04.982354 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.003093 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.023579 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.043169 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.063078 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.082246 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.102797 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.121852 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.163260 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfb47\" (UniqueName: \"kubernetes.io/projected/39f07a80-8096-439f-a3b5-b4dca2842b90-kube-api-access-qfb47\") pod \"openshift-config-operator-7777fb866f-76fgg\" (UID: \"39f07a80-8096-439f-a3b5-b4dca2842b90\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76fgg" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.182036 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqrdv\" (UniqueName: \"kubernetes.io/projected/e0c14e43-fd09-45b2-9820-abb38a0e7761-kube-api-access-dqrdv\") pod \"kube-storage-version-migrator-operator-b67b599dd-6gqzf\" (UID: \"e0c14e43-fd09-45b2-9820-abb38a0e7761\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6gqzf" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.205085 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjnxd\" (UniqueName: \"kubernetes.io/projected/f354b96e-de50-4ad1-8a82-60d570b1c6b5-kube-api-access-jjnxd\") pod \"apiserver-7bbb656c7d-hh7hq\" (UID: \"f354b96e-de50-4ad1-8a82-60d570b1c6b5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.220052 4745 request.go:700] Waited for 1.857889767s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console-operator/serviceaccounts/console-operator/token Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.225764 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zgdj\" (UniqueName: \"kubernetes.io/projected/f9365b5c-a2b4-464c-aef6-afd71c5d3427-kube-api-access-5zgdj\") pod \"route-controller-manager-6576b87f9c-nwvhl\" (UID: \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.225968 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-76fgg" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.242761 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6gqzf" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.249555 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpmsh\" (UniqueName: \"kubernetes.io/projected/420a8490-c5b6-4d9b-a716-efc526b5f4a5-kube-api-access-jpmsh\") pod \"console-operator-58897d9998-7v4cv\" (UID: \"420a8490-c5b6-4d9b-a716-efc526b5f4a5\") " pod="openshift-console-operator/console-operator-58897d9998-7v4cv" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.259652 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c157d5c1-d9e5-4b0d-94b7-432bb85ffa10-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qs225\" (UID: \"c157d5c1-d9e5-4b0d-94b7-432bb85ffa10\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.281015 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfw7c\" (UniqueName: \"kubernetes.io/projected/25f468fc-1ca1-461d-a077-2150d42bcafe-kube-api-access-mfw7c\") pod \"machine-config-operator-74547568cd-rvnrp\" (UID: \"25f468fc-1ca1-461d-a077-2150d42bcafe\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.298988 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkdqw\" (UniqueName: \"kubernetes.io/projected/f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9-kube-api-access-pkdqw\") pod \"router-default-5444994796-5ftts\" (UID: \"f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9\") " pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.324411 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbqrq\" (UniqueName: \"kubernetes.io/projected/800c6118-3c93-4dd4-b7e4-17f23b82de1d-kube-api-access-lbqrq\") pod \"machine-api-operator-5694c8668f-s8wbg\" (UID: \"800c6118-3c93-4dd4-b7e4-17f23b82de1d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.344838 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lfpb\" (UniqueName: \"kubernetes.io/projected/7ddc713d-4db8-4e77-acd2-0936a8e7bcb1-kube-api-access-8lfpb\") pod \"openshift-controller-manager-operator-756b6f6bc6-dfjd5\" (UID: \"7ddc713d-4db8-4e77-acd2-0936a8e7bcb1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dfjd5" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.352780 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.364752 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.374722 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfxkz\" (UniqueName: \"kubernetes.io/projected/f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75-kube-api-access-zfxkz\") pod \"cluster-image-registry-operator-dc59b4c8b-fk8ff\" (UID: \"f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.387040 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.390355 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-7v4cv" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.393943 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67d48\" (UniqueName: \"kubernetes.io/projected/88252921-5e6b-4e51-95e8-59dd9dd891b6-kube-api-access-67d48\") pod \"authentication-operator-69f744f599-bgmxs\" (UID: \"88252921-5e6b-4e51-95e8-59dd9dd891b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.412908 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6s46\" (UniqueName: \"kubernetes.io/projected/8f35439c-e788-4e19-8054-6a365d28a76d-kube-api-access-z6s46\") pod \"cluster-samples-operator-665b6dd947-kcwwb\" (UID: \"8f35439c-e788-4e19-8054-6a365d28a76d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kcwwb" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.415020 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fk8ff\" (UID: \"f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.423294 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.429489 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kcwwb" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.447053 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk4gk\" (UniqueName: \"kubernetes.io/projected/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-kube-api-access-tk4gk\") pod \"oauth-openshift-558db77b4-98dn4\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.461255 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-76fgg"] Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.465377 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ntfd\" (UniqueName: \"kubernetes.io/projected/d159f38f-c259-44cf-a5f9-8c4911d7864c-kube-api-access-4ntfd\") pod \"openshift-apiserver-operator-796bbdcf4f-dtjt9\" (UID: \"d159f38f-c259-44cf-a5f9-8c4911d7864c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dtjt9" Oct 10 13:20:05 crc kubenswrapper[4745]: W1010 13:20:05.480762 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39f07a80_8096_439f_a3b5_b4dca2842b90.slice/crio-db9b423ad8eb9ced3db08639d043039eecb7d178c07141905d5aac0fb23b183c WatchSource:0}: Error finding container db9b423ad8eb9ced3db08639d043039eecb7d178c07141905d5aac0fb23b183c: Status 404 returned error can't find the container with id db9b423ad8eb9ced3db08639d043039eecb7d178c07141905d5aac0fb23b183c Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.482981 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hskss\" (UniqueName: \"kubernetes.io/projected/60290e46-da36-4e07-a1db-edc753eb0788-kube-api-access-hskss\") pod \"machine-approver-56656f9798-pg69q\" (UID: \"60290e46-da36-4e07-a1db-edc753eb0788\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.502158 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8nrx\" (UniqueName: \"kubernetes.io/projected/0b152ecd-75c1-4677-8f8a-c27ff65d76d9-kube-api-access-m8nrx\") pod \"migrator-59844c95c7-hqxgq\" (UID: \"0b152ecd-75c1-4677-8f8a-c27ff65d76d9\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hqxgq" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.512208 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hqxgq" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.515129 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6gqzf"] Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.518598 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.521203 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqbfm\" (UniqueName: \"kubernetes.io/projected/0088430d-3a96-46ea-a149-7c59eb24d706-kube-api-access-gqbfm\") pod \"controller-manager-879f6c89f-9q9xb\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.536780 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:05 crc kubenswrapper[4745]: W1010 13:20:05.541641 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0c14e43_fd09_45b2_9820_abb38a0e7761.slice/crio-cf22d635783f2d1a9bf5c42010b32118908e88b9f929d463311e65c7dcde0915 WatchSource:0}: Error finding container cf22d635783f2d1a9bf5c42010b32118908e88b9f929d463311e65c7dcde0915: Status 404 returned error can't find the container with id cf22d635783f2d1a9bf5c42010b32118908e88b9f929d463311e65c7dcde0915 Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.546525 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptc8p\" (UniqueName: \"kubernetes.io/projected/247274a0-c2f0-4150-ad4f-cf39c2003a60-kube-api-access-ptc8p\") pod \"downloads-7954f5f757-7jjxn\" (UID: \"247274a0-c2f0-4150-ad4f-cf39c2003a60\") " pod="openshift-console/downloads-7954f5f757-7jjxn" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.547834 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dfjd5" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.557670 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.569119 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzrc9\" (UniqueName: \"kubernetes.io/projected/c157d5c1-d9e5-4b0d-94b7-432bb85ffa10-kube-api-access-rzrc9\") pod \"ingress-operator-5b745b69d9-qs225\" (UID: \"c157d5c1-d9e5-4b0d-94b7-432bb85ffa10\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.571771 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.572053 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq"] Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.576908 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9lsm\" (UniqueName: \"kubernetes.io/projected/839da6a0-320a-45d1-8a6c-be916ee27421-kube-api-access-t9lsm\") pod \"control-plane-machine-set-operator-78cbb6b69f-fqb4s\" (UID: \"839da6a0-320a-45d1-8a6c-be916ee27421\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fqb4s" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.593588 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.601764 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p524m\" (UniqueName: \"kubernetes.io/projected/a288a2bb-c87a-464b-b2fe-d2cab9a5df70-kube-api-access-p524m\") pod \"apiserver-76f77b778f-72w5j\" (UID: \"a288a2bb-c87a-464b-b2fe-d2cab9a5df70\") " pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.621570 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.641247 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.675876 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.682956 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 10 13:20:05 crc kubenswrapper[4745]: W1010 13:20:05.688694 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60290e46_da36_4e07_a1db_edc753eb0788.slice/crio-2240e0012d25c1bea60bf09a57c92e5d9537f98ad63628788189ac1f4138d290 WatchSource:0}: Error finding container 2240e0012d25c1bea60bf09a57c92e5d9537f98ad63628788189ac1f4138d290: Status 404 returned error can't find the container with id 2240e0012d25c1bea60bf09a57c92e5d9537f98ad63628788189ac1f4138d290 Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.691310 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-registry-tls\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.691340 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-registry-certificates\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.691366 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-bound-sa-token\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.691412 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.691430 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wwq8\" (UniqueName: \"kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-kube-api-access-7wwq8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.691457 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.691480 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.691521 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-trusted-ca\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: E1010 13:20:05.691762 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:06.191745578 +0000 UTC m=+120.089402341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.699626 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dtjt9" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.701651 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.705184 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fqb4s" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.722501 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.741634 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.741740 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.748459 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-7jjxn" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.755651 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.795598 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.795879 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f9346bb5-09f3-4539-b678-198f2e8da9a4-etcd-ca\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.795905 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97baf9cf-d52a-421e-b6b5-5996d71adf41-config\") pod \"service-ca-operator-777779d784-vrllq\" (UID: \"97baf9cf-d52a-421e-b6b5-5996d71adf41\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vrllq" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.795953 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc9kw\" (UniqueName: \"kubernetes.io/projected/f9346bb5-09f3-4539-b678-198f2e8da9a4-kube-api-access-qc9kw\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.795995 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/340979c3-ead3-4393-9aeb-3c6957ccdf80-profile-collector-cert\") pod \"olm-operator-6b444d44fb-skrn4\" (UID: \"340979c3-ead3-4393-9aeb-3c6957ccdf80\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796033 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65704ac7-fe35-4513-85b3-d512314162e8-config-volume\") pod \"dns-default-gffs7\" (UID: \"65704ac7-fe35-4513-85b3-d512314162e8\") " pod="openshift-dns/dns-default-gffs7" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796052 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/870f50c7-a485-45ae-8911-eaa1aa314e44-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-j6gkw\" (UID: \"870f50c7-a485-45ae-8911-eaa1aa314e44\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j6gkw" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796068 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-service-ca\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796083 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-oauth-config\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796096 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn2r5\" (UniqueName: \"kubernetes.io/projected/65704ac7-fe35-4513-85b3-d512314162e8-kube-api-access-tn2r5\") pod \"dns-default-gffs7\" (UID: \"65704ac7-fe35-4513-85b3-d512314162e8\") " pod="openshift-dns/dns-default-gffs7" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796112 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97baf9cf-d52a-421e-b6b5-5996d71adf41-serving-cert\") pod \"service-ca-operator-777779d784-vrllq\" (UID: \"97baf9cf-d52a-421e-b6b5-5996d71adf41\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vrllq" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796146 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ldwf\" (UniqueName: \"kubernetes.io/projected/4c8c7598-c1b9-4201-9f79-21c0c312f26a-kube-api-access-9ldwf\") pod \"packageserver-d55dfcdfc-bmlmb\" (UID: \"4c8c7598-c1b9-4201-9f79-21c0c312f26a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796169 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdzm7\" (UniqueName: \"kubernetes.io/projected/70186f6c-521c-49d0-a772-025dfb2deb76-kube-api-access-rdzm7\") pod \"catalog-operator-68c6474976-tnbzj\" (UID: \"70186f6c-521c-49d0-a772-025dfb2deb76\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796186 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7szt\" (UniqueName: \"kubernetes.io/projected/c0a536ca-a2c3-4cbf-9a0d-54fb68f80510-kube-api-access-g7szt\") pod \"dns-operator-744455d44c-ncgfh\" (UID: \"c0a536ca-a2c3-4cbf-9a0d-54fb68f80510\") " pod="openshift-dns-operator/dns-operator-744455d44c-ncgfh" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796213 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9346bb5-09f3-4539-b678-198f2e8da9a4-config\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796239 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c0a536ca-a2c3-4cbf-9a0d-54fb68f80510-metrics-tls\") pod \"dns-operator-744455d44c-ncgfh\" (UID: \"c0a536ca-a2c3-4cbf-9a0d-54fb68f80510\") " pod="openshift-dns-operator/dns-operator-744455d44c-ncgfh" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796265 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08808841-2872-4cab-8ccd-208e1d29e3b0-config\") pod \"kube-apiserver-operator-766d6c64bb-lwl2w\" (UID: \"08808841-2872-4cab-8ccd-208e1d29e3b0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lwl2w" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796280 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1bca2f23-df4e-4144-8253-9ad9b08df710-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-kth6t\" (UID: \"1bca2f23-df4e-4144-8253-9ad9b08df710\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-kth6t" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796296 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f9346bb5-09f3-4539-b678-198f2e8da9a4-etcd-client\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796337 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796356 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wwq8\" (UniqueName: \"kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-kube-api-access-7wwq8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: E1010 13:20:05.796371 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:06.296351218 +0000 UTC m=+120.194007981 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796411 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfa4b474-6bd6-4cc3-910e-d1199af5f7c6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-npfv4\" (UID: \"cfa4b474-6bd6-4cc3-910e-d1199af5f7c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-npfv4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796521 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-config-volume\") pod \"collect-profiles-29335035-mtp5w\" (UID: \"5cb8607d-1cbd-4b3a-9b06-f2db94952d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796568 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796590 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/089f446f-66e5-4e2b-99cd-9d72f05c0990-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-jr8mx\" (UID: \"089f446f-66e5-4e2b-99cd-9d72f05c0990\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jr8mx" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796617 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/08808841-2872-4cab-8ccd-208e1d29e3b0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-lwl2w\" (UID: \"08808841-2872-4cab-8ccd-208e1d29e3b0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lwl2w" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796633 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9346bb5-09f3-4539-b678-198f2e8da9a4-serving-cert\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796659 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/4c8c7598-c1b9-4201-9f79-21c0c312f26a-tmpfs\") pod \"packageserver-d55dfcdfc-bmlmb\" (UID: \"4c8c7598-c1b9-4201-9f79-21c0c312f26a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.796699 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dc27t\" (UID: \"7bff784a-c53a-4ff1-8507-e45b2eab9f1f\") " pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797029 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gw6t\" (UniqueName: \"kubernetes.io/projected/a8a4af69-070d-48f1-8f56-1466adc9efb6-kube-api-access-9gw6t\") pod \"machine-config-controller-84d6567774-mk9cf\" (UID: \"a8a4af69-070d-48f1-8f56-1466adc9efb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797093 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/70186f6c-521c-49d0-a772-025dfb2deb76-profile-collector-cert\") pod \"catalog-operator-68c6474976-tnbzj\" (UID: \"70186f6c-521c-49d0-a772-025dfb2deb76\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797192 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-trusted-ca\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797209 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f7bj\" (UniqueName: \"kubernetes.io/projected/340979c3-ead3-4393-9aeb-3c6957ccdf80-kube-api-access-9f7bj\") pod \"olm-operator-6b444d44fb-skrn4\" (UID: \"340979c3-ead3-4393-9aeb-3c6957ccdf80\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797255 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/03110c82-b25a-4ed8-a9ae-b59bd8181baf-mountpoint-dir\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797290 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czm9p\" (UniqueName: \"kubernetes.io/projected/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-kube-api-access-czm9p\") pod \"marketplace-operator-79b997595-dc27t\" (UID: \"7bff784a-c53a-4ff1-8507-e45b2eab9f1f\") " pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797305 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-secret-volume\") pod \"collect-profiles-29335035-mtp5w\" (UID: \"5cb8607d-1cbd-4b3a-9b06-f2db94952d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797330 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef-certs\") pod \"machine-config-server-pdt5f\" (UID: \"e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef\") " pod="openshift-machine-config-operator/machine-config-server-pdt5f" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797344 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg9zp\" (UniqueName: \"kubernetes.io/projected/870f50c7-a485-45ae-8911-eaa1aa314e44-kube-api-access-hg9zp\") pod \"package-server-manager-789f6589d5-j6gkw\" (UID: \"870f50c7-a485-45ae-8911-eaa1aa314e44\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j6gkw" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797360 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8a4af69-070d-48f1-8f56-1466adc9efb6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mk9cf\" (UID: \"a8a4af69-070d-48f1-8f56-1466adc9efb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797415 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-registry-tls\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797431 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/340979c3-ead3-4393-9aeb-3c6957ccdf80-srv-cert\") pod \"olm-operator-6b444d44fb-skrn4\" (UID: \"340979c3-ead3-4393-9aeb-3c6957ccdf80\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797467 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-registry-certificates\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797482 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-oauth-serving-cert\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797518 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbxbd\" (UniqueName: \"kubernetes.io/projected/1bca2f23-df4e-4144-8253-9ad9b08df710-kube-api-access-nbxbd\") pod \"multus-admission-controller-857f4d67dd-kth6t\" (UID: \"1bca2f23-df4e-4144-8253-9ad9b08df710\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-kth6t" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797533 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd67c\" (UniqueName: \"kubernetes.io/projected/83aaa6eb-34bb-4755-a8af-bd5925aef46c-kube-api-access-cd67c\") pod \"service-ca-9c57cc56f-kgpj4\" (UID: \"83aaa6eb-34bb-4755-a8af-bd5925aef46c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kgpj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797557 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6cw5\" (UniqueName: \"kubernetes.io/projected/e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef-kube-api-access-b6cw5\") pod \"machine-config-server-pdt5f\" (UID: \"e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef\") " pod="openshift-machine-config-operator/machine-config-server-pdt5f" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797589 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/03110c82-b25a-4ed8-a9ae-b59bd8181baf-registration-dir\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797603 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cfa4b474-6bd6-4cc3-910e-d1199af5f7c6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-npfv4\" (UID: \"cfa4b474-6bd6-4cc3-910e-d1199af5f7c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-npfv4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797627 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-bound-sa-token\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797653 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfa4b474-6bd6-4cc3-910e-d1199af5f7c6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-npfv4\" (UID: \"cfa4b474-6bd6-4cc3-910e-d1199af5f7c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-npfv4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797698 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4c8c7598-c1b9-4201-9f79-21c0c312f26a-webhook-cert\") pod \"packageserver-d55dfcdfc-bmlmb\" (UID: \"4c8c7598-c1b9-4201-9f79-21c0c312f26a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797776 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/089f446f-66e5-4e2b-99cd-9d72f05c0990-config\") pod \"kube-controller-manager-operator-78b949d7b-jr8mx\" (UID: \"089f446f-66e5-4e2b-99cd-9d72f05c0990\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jr8mx" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797809 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htmd2\" (UniqueName: \"kubernetes.io/projected/e8bcad89-0f86-4a07-9c81-aa8cac743319-kube-api-access-htmd2\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797845 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f9346bb5-09f3-4539-b678-198f2e8da9a4-etcd-service-ca\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797862 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/70186f6c-521c-49d0-a772-025dfb2deb76-srv-cert\") pod \"catalog-operator-68c6474976-tnbzj\" (UID: \"70186f6c-521c-49d0-a772-025dfb2deb76\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797878 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8a4af69-070d-48f1-8f56-1466adc9efb6-proxy-tls\") pod \"machine-config-controller-84d6567774-mk9cf\" (UID: \"a8a4af69-070d-48f1-8f56-1466adc9efb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797896 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-serving-cert\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797926 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08808841-2872-4cab-8ccd-208e1d29e3b0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-lwl2w\" (UID: \"08808841-2872-4cab-8ccd-208e1d29e3b0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lwl2w" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.797962 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef-node-bootstrap-token\") pod \"machine-config-server-pdt5f\" (UID: \"e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef\") " pod="openshift-machine-config-operator/machine-config-server-pdt5f" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.798701 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.798778 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/03110c82-b25a-4ed8-a9ae-b59bd8181baf-plugins-dir\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.798805 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-config\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.798858 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4c8c7598-c1b9-4201-9f79-21c0c312f26a-apiservice-cert\") pod \"packageserver-d55dfcdfc-bmlmb\" (UID: \"4c8c7598-c1b9-4201-9f79-21c0c312f26a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.798878 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/469f0496-8d18-43a5-b25f-a9450f92fac7-cert\") pod \"ingress-canary-zdbqc\" (UID: \"469f0496-8d18-43a5-b25f-a9450f92fac7\") " pod="openshift-ingress-canary/ingress-canary-zdbqc" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.798894 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m95l\" (UniqueName: \"kubernetes.io/projected/469f0496-8d18-43a5-b25f-a9450f92fac7-kube-api-access-9m95l\") pod \"ingress-canary-zdbqc\" (UID: \"469f0496-8d18-43a5-b25f-a9450f92fac7\") " pod="openshift-ingress-canary/ingress-canary-zdbqc" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.798946 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/03110c82-b25a-4ed8-a9ae-b59bd8181baf-csi-data-dir\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.798965 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/83aaa6eb-34bb-4755-a8af-bd5925aef46c-signing-cabundle\") pod \"service-ca-9c57cc56f-kgpj4\" (UID: \"83aaa6eb-34bb-4755-a8af-bd5925aef46c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kgpj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.799020 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/03110c82-b25a-4ed8-a9ae-b59bd8181baf-socket-dir\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.799038 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dc27t\" (UID: \"7bff784a-c53a-4ff1-8507-e45b2eab9f1f\") " pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.799067 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-trusted-ca-bundle\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.799112 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbwsw\" (UniqueName: \"kubernetes.io/projected/03110c82-b25a-4ed8-a9ae-b59bd8181baf-kube-api-access-dbwsw\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.799132 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxg27\" (UniqueName: \"kubernetes.io/projected/97baf9cf-d52a-421e-b6b5-5996d71adf41-kube-api-access-kxg27\") pod \"service-ca-operator-777779d784-vrllq\" (UID: \"97baf9cf-d52a-421e-b6b5-5996d71adf41\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vrllq" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.799230 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/65704ac7-fe35-4513-85b3-d512314162e8-metrics-tls\") pod \"dns-default-gffs7\" (UID: \"65704ac7-fe35-4513-85b3-d512314162e8\") " pod="openshift-dns/dns-default-gffs7" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.799247 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/83aaa6eb-34bb-4755-a8af-bd5925aef46c-signing-key\") pod \"service-ca-9c57cc56f-kgpj4\" (UID: \"83aaa6eb-34bb-4755-a8af-bd5925aef46c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kgpj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.799266 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9nxm\" (UniqueName: \"kubernetes.io/projected/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-kube-api-access-l9nxm\") pod \"collect-profiles-29335035-mtp5w\" (UID: \"5cb8607d-1cbd-4b3a-9b06-f2db94952d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.799285 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/089f446f-66e5-4e2b-99cd-9d72f05c0990-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-jr8mx\" (UID: \"089f446f-66e5-4e2b-99cd-9d72f05c0990\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jr8mx" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.808723 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.810539 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-registry-certificates\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: E1010 13:20:05.811122 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:06.311105613 +0000 UTC m=+120.208762376 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.812302 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-registry-tls\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.820357 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.823066 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-trusted-ca\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.840934 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wwq8\" (UniqueName: \"kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-kube-api-access-7wwq8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.876489 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-bound-sa-token\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.894100 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl"] Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.896515 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7v4cv"] Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.896572 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp"] Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.900420 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:05 crc kubenswrapper[4745]: E1010 13:20:05.900599 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:06.400581838 +0000 UTC m=+120.298238601 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.900757 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef-node-bootstrap-token\") pod \"machine-config-server-pdt5f\" (UID: \"e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef\") " pod="openshift-machine-config-operator/machine-config-server-pdt5f" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.900793 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/03110c82-b25a-4ed8-a9ae-b59bd8181baf-plugins-dir\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.900964 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-config\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.901451 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/03110c82-b25a-4ed8-a9ae-b59bd8181baf-plugins-dir\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.902154 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-config\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.902638 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4c8c7598-c1b9-4201-9f79-21c0c312f26a-apiservice-cert\") pod \"packageserver-d55dfcdfc-bmlmb\" (UID: \"4c8c7598-c1b9-4201-9f79-21c0c312f26a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.902685 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/469f0496-8d18-43a5-b25f-a9450f92fac7-cert\") pod \"ingress-canary-zdbqc\" (UID: \"469f0496-8d18-43a5-b25f-a9450f92fac7\") " pod="openshift-ingress-canary/ingress-canary-zdbqc" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.902706 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m95l\" (UniqueName: \"kubernetes.io/projected/469f0496-8d18-43a5-b25f-a9450f92fac7-kube-api-access-9m95l\") pod \"ingress-canary-zdbqc\" (UID: \"469f0496-8d18-43a5-b25f-a9450f92fac7\") " pod="openshift-ingress-canary/ingress-canary-zdbqc" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903100 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/03110c82-b25a-4ed8-a9ae-b59bd8181baf-socket-dir\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903113 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/03110c82-b25a-4ed8-a9ae-b59bd8181baf-socket-dir\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903134 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/03110c82-b25a-4ed8-a9ae-b59bd8181baf-csi-data-dir\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903157 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/83aaa6eb-34bb-4755-a8af-bd5925aef46c-signing-cabundle\") pod \"service-ca-9c57cc56f-kgpj4\" (UID: \"83aaa6eb-34bb-4755-a8af-bd5925aef46c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kgpj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903178 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dc27t\" (UID: \"7bff784a-c53a-4ff1-8507-e45b2eab9f1f\") " pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903197 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/03110c82-b25a-4ed8-a9ae-b59bd8181baf-csi-data-dir\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903201 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-trusted-ca-bundle\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903272 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbwsw\" (UniqueName: \"kubernetes.io/projected/03110c82-b25a-4ed8-a9ae-b59bd8181baf-kube-api-access-dbwsw\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903292 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxg27\" (UniqueName: \"kubernetes.io/projected/97baf9cf-d52a-421e-b6b5-5996d71adf41-kube-api-access-kxg27\") pod \"service-ca-operator-777779d784-vrllq\" (UID: \"97baf9cf-d52a-421e-b6b5-5996d71adf41\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vrllq" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903347 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/089f446f-66e5-4e2b-99cd-9d72f05c0990-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-jr8mx\" (UID: \"089f446f-66e5-4e2b-99cd-9d72f05c0990\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jr8mx" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903368 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/65704ac7-fe35-4513-85b3-d512314162e8-metrics-tls\") pod \"dns-default-gffs7\" (UID: \"65704ac7-fe35-4513-85b3-d512314162e8\") " pod="openshift-dns/dns-default-gffs7" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903388 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/83aaa6eb-34bb-4755-a8af-bd5925aef46c-signing-key\") pod \"service-ca-9c57cc56f-kgpj4\" (UID: \"83aaa6eb-34bb-4755-a8af-bd5925aef46c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kgpj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903408 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9nxm\" (UniqueName: \"kubernetes.io/projected/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-kube-api-access-l9nxm\") pod \"collect-profiles-29335035-mtp5w\" (UID: \"5cb8607d-1cbd-4b3a-9b06-f2db94952d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903442 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f9346bb5-09f3-4539-b678-198f2e8da9a4-etcd-ca\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903463 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97baf9cf-d52a-421e-b6b5-5996d71adf41-config\") pod \"service-ca-operator-777779d784-vrllq\" (UID: \"97baf9cf-d52a-421e-b6b5-5996d71adf41\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vrllq" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903483 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc9kw\" (UniqueName: \"kubernetes.io/projected/f9346bb5-09f3-4539-b678-198f2e8da9a4-kube-api-access-qc9kw\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903506 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/340979c3-ead3-4393-9aeb-3c6957ccdf80-profile-collector-cert\") pod \"olm-operator-6b444d44fb-skrn4\" (UID: \"340979c3-ead3-4393-9aeb-3c6957ccdf80\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903527 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65704ac7-fe35-4513-85b3-d512314162e8-config-volume\") pod \"dns-default-gffs7\" (UID: \"65704ac7-fe35-4513-85b3-d512314162e8\") " pod="openshift-dns/dns-default-gffs7" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903551 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/870f50c7-a485-45ae-8911-eaa1aa314e44-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-j6gkw\" (UID: \"870f50c7-a485-45ae-8911-eaa1aa314e44\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j6gkw" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903571 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-service-ca\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903590 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn2r5\" (UniqueName: \"kubernetes.io/projected/65704ac7-fe35-4513-85b3-d512314162e8-kube-api-access-tn2r5\") pod \"dns-default-gffs7\" (UID: \"65704ac7-fe35-4513-85b3-d512314162e8\") " pod="openshift-dns/dns-default-gffs7" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903608 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-oauth-config\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903627 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97baf9cf-d52a-421e-b6b5-5996d71adf41-serving-cert\") pod \"service-ca-operator-777779d784-vrllq\" (UID: \"97baf9cf-d52a-421e-b6b5-5996d71adf41\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vrllq" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903648 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ldwf\" (UniqueName: \"kubernetes.io/projected/4c8c7598-c1b9-4201-9f79-21c0c312f26a-kube-api-access-9ldwf\") pod \"packageserver-d55dfcdfc-bmlmb\" (UID: \"4c8c7598-c1b9-4201-9f79-21c0c312f26a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903669 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdzm7\" (UniqueName: \"kubernetes.io/projected/70186f6c-521c-49d0-a772-025dfb2deb76-kube-api-access-rdzm7\") pod \"catalog-operator-68c6474976-tnbzj\" (UID: \"70186f6c-521c-49d0-a772-025dfb2deb76\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903690 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7szt\" (UniqueName: \"kubernetes.io/projected/c0a536ca-a2c3-4cbf-9a0d-54fb68f80510-kube-api-access-g7szt\") pod \"dns-operator-744455d44c-ncgfh\" (UID: \"c0a536ca-a2c3-4cbf-9a0d-54fb68f80510\") " pod="openshift-dns-operator/dns-operator-744455d44c-ncgfh" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903721 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c0a536ca-a2c3-4cbf-9a0d-54fb68f80510-metrics-tls\") pod \"dns-operator-744455d44c-ncgfh\" (UID: \"c0a536ca-a2c3-4cbf-9a0d-54fb68f80510\") " pod="openshift-dns-operator/dns-operator-744455d44c-ncgfh" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903758 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9346bb5-09f3-4539-b678-198f2e8da9a4-config\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903780 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08808841-2872-4cab-8ccd-208e1d29e3b0-config\") pod \"kube-apiserver-operator-766d6c64bb-lwl2w\" (UID: \"08808841-2872-4cab-8ccd-208e1d29e3b0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lwl2w" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903800 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1bca2f23-df4e-4144-8253-9ad9b08df710-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-kth6t\" (UID: \"1bca2f23-df4e-4144-8253-9ad9b08df710\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-kth6t" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903821 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f9346bb5-09f3-4539-b678-198f2e8da9a4-etcd-client\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903846 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfa4b474-6bd6-4cc3-910e-d1199af5f7c6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-npfv4\" (UID: \"cfa4b474-6bd6-4cc3-910e-d1199af5f7c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-npfv4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903872 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-config-volume\") pod \"collect-profiles-29335035-mtp5w\" (UID: \"5cb8607d-1cbd-4b3a-9b06-f2db94952d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903898 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903922 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/089f446f-66e5-4e2b-99cd-9d72f05c0990-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-jr8mx\" (UID: \"089f446f-66e5-4e2b-99cd-9d72f05c0990\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jr8mx" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903944 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/08808841-2872-4cab-8ccd-208e1d29e3b0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-lwl2w\" (UID: \"08808841-2872-4cab-8ccd-208e1d29e3b0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lwl2w" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903963 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9346bb5-09f3-4539-b678-198f2e8da9a4-serving-cert\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.903985 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/4c8c7598-c1b9-4201-9f79-21c0c312f26a-tmpfs\") pod \"packageserver-d55dfcdfc-bmlmb\" (UID: \"4c8c7598-c1b9-4201-9f79-21c0c312f26a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.904007 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dc27t\" (UID: \"7bff784a-c53a-4ff1-8507-e45b2eab9f1f\") " pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.904030 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gw6t\" (UniqueName: \"kubernetes.io/projected/a8a4af69-070d-48f1-8f56-1466adc9efb6-kube-api-access-9gw6t\") pod \"machine-config-controller-84d6567774-mk9cf\" (UID: \"a8a4af69-070d-48f1-8f56-1466adc9efb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.904053 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/70186f6c-521c-49d0-a772-025dfb2deb76-profile-collector-cert\") pod \"catalog-operator-68c6474976-tnbzj\" (UID: \"70186f6c-521c-49d0-a772-025dfb2deb76\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.904086 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f7bj\" (UniqueName: \"kubernetes.io/projected/340979c3-ead3-4393-9aeb-3c6957ccdf80-kube-api-access-9f7bj\") pod \"olm-operator-6b444d44fb-skrn4\" (UID: \"340979c3-ead3-4393-9aeb-3c6957ccdf80\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.904109 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/03110c82-b25a-4ed8-a9ae-b59bd8181baf-mountpoint-dir\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.904132 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czm9p\" (UniqueName: \"kubernetes.io/projected/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-kube-api-access-czm9p\") pod \"marketplace-operator-79b997595-dc27t\" (UID: \"7bff784a-c53a-4ff1-8507-e45b2eab9f1f\") " pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.904151 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef-certs\") pod \"machine-config-server-pdt5f\" (UID: \"e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef\") " pod="openshift-machine-config-operator/machine-config-server-pdt5f" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.904171 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-secret-volume\") pod \"collect-profiles-29335035-mtp5w\" (UID: \"5cb8607d-1cbd-4b3a-9b06-f2db94952d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.904191 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg9zp\" (UniqueName: \"kubernetes.io/projected/870f50c7-a485-45ae-8911-eaa1aa314e44-kube-api-access-hg9zp\") pod \"package-server-manager-789f6589d5-j6gkw\" (UID: \"870f50c7-a485-45ae-8911-eaa1aa314e44\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j6gkw" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.904213 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8a4af69-070d-48f1-8f56-1466adc9efb6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mk9cf\" (UID: \"a8a4af69-070d-48f1-8f56-1466adc9efb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.904235 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/340979c3-ead3-4393-9aeb-3c6957ccdf80-srv-cert\") pod \"olm-operator-6b444d44fb-skrn4\" (UID: \"340979c3-ead3-4393-9aeb-3c6957ccdf80\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.904257 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-oauth-serving-cert\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.904307 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbxbd\" (UniqueName: \"kubernetes.io/projected/1bca2f23-df4e-4144-8253-9ad9b08df710-kube-api-access-nbxbd\") pod \"multus-admission-controller-857f4d67dd-kth6t\" (UID: \"1bca2f23-df4e-4144-8253-9ad9b08df710\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-kth6t" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.904331 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd67c\" (UniqueName: \"kubernetes.io/projected/83aaa6eb-34bb-4755-a8af-bd5925aef46c-kube-api-access-cd67c\") pod \"service-ca-9c57cc56f-kgpj4\" (UID: \"83aaa6eb-34bb-4755-a8af-bd5925aef46c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kgpj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.904353 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6cw5\" (UniqueName: \"kubernetes.io/projected/e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef-kube-api-access-b6cw5\") pod \"machine-config-server-pdt5f\" (UID: \"e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef\") " pod="openshift-machine-config-operator/machine-config-server-pdt5f" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.904375 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/03110c82-b25a-4ed8-a9ae-b59bd8181baf-registration-dir\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.904945 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/83aaa6eb-34bb-4755-a8af-bd5925aef46c-signing-cabundle\") pod \"service-ca-9c57cc56f-kgpj4\" (UID: \"83aaa6eb-34bb-4755-a8af-bd5925aef46c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kgpj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.905795 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cfa4b474-6bd6-4cc3-910e-d1199af5f7c6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-npfv4\" (UID: \"cfa4b474-6bd6-4cc3-910e-d1199af5f7c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-npfv4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.905869 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfa4b474-6bd6-4cc3-910e-d1199af5f7c6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-npfv4\" (UID: \"cfa4b474-6bd6-4cc3-910e-d1199af5f7c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-npfv4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.905894 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4c8c7598-c1b9-4201-9f79-21c0c312f26a-webhook-cert\") pod \"packageserver-d55dfcdfc-bmlmb\" (UID: \"4c8c7598-c1b9-4201-9f79-21c0c312f26a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.905946 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/089f446f-66e5-4e2b-99cd-9d72f05c0990-config\") pod \"kube-controller-manager-operator-78b949d7b-jr8mx\" (UID: \"089f446f-66e5-4e2b-99cd-9d72f05c0990\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jr8mx" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.905976 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htmd2\" (UniqueName: \"kubernetes.io/projected/e8bcad89-0f86-4a07-9c81-aa8cac743319-kube-api-access-htmd2\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.906001 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f9346bb5-09f3-4539-b678-198f2e8da9a4-etcd-service-ca\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.906023 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/70186f6c-521c-49d0-a772-025dfb2deb76-srv-cert\") pod \"catalog-operator-68c6474976-tnbzj\" (UID: \"70186f6c-521c-49d0-a772-025dfb2deb76\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.906046 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8a4af69-070d-48f1-8f56-1466adc9efb6-proxy-tls\") pod \"machine-config-controller-84d6567774-mk9cf\" (UID: \"a8a4af69-070d-48f1-8f56-1466adc9efb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.906064 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-serving-cert\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.906085 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08808841-2872-4cab-8ccd-208e1d29e3b0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-lwl2w\" (UID: \"08808841-2872-4cab-8ccd-208e1d29e3b0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lwl2w" Oct 10 13:20:05 crc kubenswrapper[4745]: E1010 13:20:05.907100 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:06.407083081 +0000 UTC m=+120.304739844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.908395 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dfjd5"] Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.909593 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dc27t\" (UID: \"7bff784a-c53a-4ff1-8507-e45b2eab9f1f\") " pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.909677 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/469f0496-8d18-43a5-b25f-a9450f92fac7-cert\") pod \"ingress-canary-zdbqc\" (UID: \"469f0496-8d18-43a5-b25f-a9450f92fac7\") " pod="openshift-ingress-canary/ingress-canary-zdbqc" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.910959 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08808841-2872-4cab-8ccd-208e1d29e3b0-config\") pod \"kube-apiserver-operator-766d6c64bb-lwl2w\" (UID: \"08808841-2872-4cab-8ccd-208e1d29e3b0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lwl2w" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.911021 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef-node-bootstrap-token\") pod \"machine-config-server-pdt5f\" (UID: \"e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef\") " pod="openshift-machine-config-operator/machine-config-server-pdt5f" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.911837 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-trusted-ca-bundle\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.911955 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9346bb5-09f3-4539-b678-198f2e8da9a4-config\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.912375 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f9346bb5-09f3-4539-b678-198f2e8da9a4-etcd-ca\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.912663 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/03110c82-b25a-4ed8-a9ae-b59bd8181baf-mountpoint-dir\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.912924 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97baf9cf-d52a-421e-b6b5-5996d71adf41-config\") pod \"service-ca-operator-777779d784-vrllq\" (UID: \"97baf9cf-d52a-421e-b6b5-5996d71adf41\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vrllq" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.913914 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08808841-2872-4cab-8ccd-208e1d29e3b0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-lwl2w\" (UID: \"08808841-2872-4cab-8ccd-208e1d29e3b0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lwl2w" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.914176 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/65704ac7-fe35-4513-85b3-d512314162e8-metrics-tls\") pod \"dns-default-gffs7\" (UID: \"65704ac7-fe35-4513-85b3-d512314162e8\") " pod="openshift-dns/dns-default-gffs7" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.914562 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-oauth-serving-cert\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.914856 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/03110c82-b25a-4ed8-a9ae-b59bd8181baf-registration-dir\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.915141 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-oauth-config\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.915187 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-service-ca\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.915250 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/4c8c7598-c1b9-4201-9f79-21c0c312f26a-tmpfs\") pod \"packageserver-d55dfcdfc-bmlmb\" (UID: \"4c8c7598-c1b9-4201-9f79-21c0c312f26a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.915862 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/83aaa6eb-34bb-4755-a8af-bd5925aef46c-signing-key\") pod \"service-ca-9c57cc56f-kgpj4\" (UID: \"83aaa6eb-34bb-4755-a8af-bd5925aef46c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kgpj4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.920616 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1bca2f23-df4e-4144-8253-9ad9b08df710-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-kth6t\" (UID: \"1bca2f23-df4e-4144-8253-9ad9b08df710\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-kth6t" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.920862 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfa4b474-6bd6-4cc3-910e-d1199af5f7c6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-npfv4\" (UID: \"cfa4b474-6bd6-4cc3-910e-d1199af5f7c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-npfv4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.921038 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef-certs\") pod \"machine-config-server-pdt5f\" (UID: \"e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef\") " pod="openshift-machine-config-operator/machine-config-server-pdt5f" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.921276 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/340979c3-ead3-4393-9aeb-3c6957ccdf80-srv-cert\") pod \"olm-operator-6b444d44fb-skrn4\" (UID: \"340979c3-ead3-4393-9aeb-3c6957ccdf80\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.922551 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-config-volume\") pod \"collect-profiles-29335035-mtp5w\" (UID: \"5cb8607d-1cbd-4b3a-9b06-f2db94952d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.922971 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f9346bb5-09f3-4539-b678-198f2e8da9a4-etcd-service-ca\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.923351 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4c8c7598-c1b9-4201-9f79-21c0c312f26a-apiservice-cert\") pod \"packageserver-d55dfcdfc-bmlmb\" (UID: \"4c8c7598-c1b9-4201-9f79-21c0c312f26a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.923396 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/089f446f-66e5-4e2b-99cd-9d72f05c0990-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-jr8mx\" (UID: \"089f446f-66e5-4e2b-99cd-9d72f05c0990\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jr8mx" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.923533 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfa4b474-6bd6-4cc3-910e-d1199af5f7c6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-npfv4\" (UID: \"cfa4b474-6bd6-4cc3-910e-d1199af5f7c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-npfv4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.924163 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4c8c7598-c1b9-4201-9f79-21c0c312f26a-webhook-cert\") pod \"packageserver-d55dfcdfc-bmlmb\" (UID: \"4c8c7598-c1b9-4201-9f79-21c0c312f26a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.926382 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/70186f6c-521c-49d0-a772-025dfb2deb76-profile-collector-cert\") pod \"catalog-operator-68c6474976-tnbzj\" (UID: \"70186f6c-521c-49d0-a772-025dfb2deb76\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.927892 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65704ac7-fe35-4513-85b3-d512314162e8-config-volume\") pod \"dns-default-gffs7\" (UID: \"65704ac7-fe35-4513-85b3-d512314162e8\") " pod="openshift-dns/dns-default-gffs7" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.928341 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/870f50c7-a485-45ae-8911-eaa1aa314e44-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-j6gkw\" (UID: \"870f50c7-a485-45ae-8911-eaa1aa314e44\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j6gkw" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.930306 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/340979c3-ead3-4393-9aeb-3c6957ccdf80-profile-collector-cert\") pod \"olm-operator-6b444d44fb-skrn4\" (UID: \"340979c3-ead3-4393-9aeb-3c6957ccdf80\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.932344 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c0a536ca-a2c3-4cbf-9a0d-54fb68f80510-metrics-tls\") pod \"dns-operator-744455d44c-ncgfh\" (UID: \"c0a536ca-a2c3-4cbf-9a0d-54fb68f80510\") " pod="openshift-dns-operator/dns-operator-744455d44c-ncgfh" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.932358 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dc27t\" (UID: \"7bff784a-c53a-4ff1-8507-e45b2eab9f1f\") " pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.935614 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8a4af69-070d-48f1-8f56-1466adc9efb6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mk9cf\" (UID: \"a8a4af69-070d-48f1-8f56-1466adc9efb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.939701 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/089f446f-66e5-4e2b-99cd-9d72f05c0990-config\") pod \"kube-controller-manager-operator-78b949d7b-jr8mx\" (UID: \"089f446f-66e5-4e2b-99cd-9d72f05c0990\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jr8mx" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.940543 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97baf9cf-d52a-421e-b6b5-5996d71adf41-serving-cert\") pod \"service-ca-operator-777779d784-vrllq\" (UID: \"97baf9cf-d52a-421e-b6b5-5996d71adf41\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vrllq" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.940619 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f9346bb5-09f3-4539-b678-198f2e8da9a4-etcd-client\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.949044 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-secret-volume\") pod \"collect-profiles-29335035-mtp5w\" (UID: \"5cb8607d-1cbd-4b3a-9b06-f2db94952d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.953163 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9346bb5-09f3-4539-b678-198f2e8da9a4-serving-cert\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.955113 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m95l\" (UniqueName: \"kubernetes.io/projected/469f0496-8d18-43a5-b25f-a9450f92fac7-kube-api-access-9m95l\") pod \"ingress-canary-zdbqc\" (UID: \"469f0496-8d18-43a5-b25f-a9450f92fac7\") " pod="openshift-ingress-canary/ingress-canary-zdbqc" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.955369 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8a4af69-070d-48f1-8f56-1466adc9efb6-proxy-tls\") pod \"machine-config-controller-84d6567774-mk9cf\" (UID: \"a8a4af69-070d-48f1-8f56-1466adc9efb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.955548 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-serving-cert\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.956420 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/70186f6c-521c-49d0-a772-025dfb2deb76-srv-cert\") pod \"catalog-operator-68c6474976-tnbzj\" (UID: \"70186f6c-521c-49d0-a772-025dfb2deb76\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.968593 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tn2r5\" (UniqueName: \"kubernetes.io/projected/65704ac7-fe35-4513-85b3-d512314162e8-kube-api-access-tn2r5\") pod \"dns-default-gffs7\" (UID: \"65704ac7-fe35-4513-85b3-d512314162e8\") " pod="openshift-dns/dns-default-gffs7" Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.981532 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbwsw\" (UniqueName: \"kubernetes.io/projected/03110c82-b25a-4ed8-a9ae-b59bd8181baf-kube-api-access-dbwsw\") pod \"csi-hostpathplugin-rwdj4\" (UID: \"03110c82-b25a-4ed8-a9ae-b59bd8181baf\") " pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:05 crc kubenswrapper[4745]: W1010 13:20:05.985034 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9365b5c_a2b4_464c_aef6_afd71c5d3427.slice/crio-6b61c5ef92d84df5681d150821515793325c11f124c3b6fd0d9b7cab2585f349 WatchSource:0}: Error finding container 6b61c5ef92d84df5681d150821515793325c11f124c3b6fd0d9b7cab2585f349: Status 404 returned error can't find the container with id 6b61c5ef92d84df5681d150821515793325c11f124c3b6fd0d9b7cab2585f349 Oct 10 13:20:05 crc kubenswrapper[4745]: I1010 13:20:05.985160 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-s8wbg"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.002698 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxg27\" (UniqueName: \"kubernetes.io/projected/97baf9cf-d52a-421e-b6b5-5996d71adf41-kube-api-access-kxg27\") pod \"service-ca-operator-777779d784-vrllq\" (UID: \"97baf9cf-d52a-421e-b6b5-5996d71adf41\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vrllq" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.007167 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:06 crc kubenswrapper[4745]: E1010 13:20:06.007530 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:06.507514984 +0000 UTC m=+120.405171747 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.008720 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kcwwb"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.016951 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/089f446f-66e5-4e2b-99cd-9d72f05c0990-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-jr8mx\" (UID: \"089f446f-66e5-4e2b-99cd-9d72f05c0990\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jr8mx" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.020441 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.036133 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-gffs7" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.044235 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-zdbqc" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.059344 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9nxm\" (UniqueName: \"kubernetes.io/projected/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-kube-api-access-l9nxm\") pod \"collect-profiles-29335035-mtp5w\" (UID: \"5cb8607d-1cbd-4b3a-9b06-f2db94952d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.082046 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ldwf\" (UniqueName: \"kubernetes.io/projected/4c8c7598-c1b9-4201-9f79-21c0c312f26a-kube-api-access-9ldwf\") pod \"packageserver-d55dfcdfc-bmlmb\" (UID: \"4c8c7598-c1b9-4201-9f79-21c0c312f26a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.109252 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdzm7\" (UniqueName: \"kubernetes.io/projected/70186f6c-521c-49d0-a772-025dfb2deb76-kube-api-access-rdzm7\") pod \"catalog-operator-68c6474976-tnbzj\" (UID: \"70186f6c-521c-49d0-a772-025dfb2deb76\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.109446 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:06 crc kubenswrapper[4745]: E1010 13:20:06.109930 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:06.609907873 +0000 UTC m=+120.507564866 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.126200 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-72w5j"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.127529 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f7bj\" (UniqueName: \"kubernetes.io/projected/340979c3-ead3-4393-9aeb-3c6957ccdf80-kube-api-access-9f7bj\") pod \"olm-operator-6b444d44fb-skrn4\" (UID: \"340979c3-ead3-4393-9aeb-3c6957ccdf80\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.143553 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7szt\" (UniqueName: \"kubernetes.io/projected/c0a536ca-a2c3-4cbf-9a0d-54fb68f80510-kube-api-access-g7szt\") pod \"dns-operator-744455d44c-ncgfh\" (UID: \"c0a536ca-a2c3-4cbf-9a0d-54fb68f80510\") " pod="openshift-dns-operator/dns-operator-744455d44c-ncgfh" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.161460 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ncgfh" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.169449 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czm9p\" (UniqueName: \"kubernetes.io/projected/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-kube-api-access-czm9p\") pod \"marketplace-operator-79b997595-dc27t\" (UID: \"7bff784a-c53a-4ff1-8507-e45b2eab9f1f\") " pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.178917 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc9kw\" (UniqueName: \"kubernetes.io/projected/f9346bb5-09f3-4539-b678-198f2e8da9a4-kube-api-access-qc9kw\") pod \"etcd-operator-b45778765-8m4nd\" (UID: \"f9346bb5-09f3-4539-b678-198f2e8da9a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.179316 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jr8mx" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.188353 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.198853 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/08808841-2872-4cab-8ccd-208e1d29e3b0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-lwl2w\" (UID: \"08808841-2872-4cab-8ccd-208e1d29e3b0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lwl2w" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.220555 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-7jjxn"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.221013 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.222020 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:06 crc kubenswrapper[4745]: E1010 13:20:06.222375 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:06.722361793 +0000 UTC m=+120.620018556 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.223064 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbxbd\" (UniqueName: \"kubernetes.io/projected/1bca2f23-df4e-4144-8253-9ad9b08df710-kube-api-access-nbxbd\") pod \"multus-admission-controller-857f4d67dd-kth6t\" (UID: \"1bca2f23-df4e-4144-8253-9ad9b08df710\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-kth6t" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.229976 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lwl2w" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.236587 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hqxgq"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.241386 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bgmxs"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.242329 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9q9xb"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.243330 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.245983 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.252779 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd67c\" (UniqueName: \"kubernetes.io/projected/83aaa6eb-34bb-4755-a8af-bd5925aef46c-kube-api-access-cd67c\") pod \"service-ca-9c57cc56f-kgpj4\" (UID: \"83aaa6eb-34bb-4755-a8af-bd5925aef46c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kgpj4" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.253105 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.260352 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6cw5\" (UniqueName: \"kubernetes.io/projected/e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef-kube-api-access-b6cw5\") pod \"machine-config-server-pdt5f\" (UID: \"e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef\") " pod="openshift-machine-config-operator/machine-config-server-pdt5f" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.262281 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vrllq" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.273144 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.277296 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-98dn4"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.287944 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.291190 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cfa4b474-6bd6-4cc3-910e-d1199af5f7c6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-npfv4\" (UID: \"cfa4b474-6bd6-4cc3-910e-d1199af5f7c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-npfv4" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.302364 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-kgpj4" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.344033 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-pdt5f" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.345633 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:06 crc kubenswrapper[4745]: E1010 13:20:06.345898 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:06.845884345 +0000 UTC m=+120.743541108 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.350798 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gw6t\" (UniqueName: \"kubernetes.io/projected/a8a4af69-070d-48f1-8f56-1466adc9efb6-kube-api-access-9gw6t\") pod \"machine-config-controller-84d6567774-mk9cf\" (UID: \"a8a4af69-070d-48f1-8f56-1466adc9efb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.369331 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg9zp\" (UniqueName: \"kubernetes.io/projected/870f50c7-a485-45ae-8911-eaa1aa314e44-kube-api-access-hg9zp\") pod \"package-server-manager-789f6589d5-j6gkw\" (UID: \"870f50c7-a485-45ae-8911-eaa1aa314e44\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j6gkw" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.402399 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qs225"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.410049 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htmd2\" (UniqueName: \"kubernetes.io/projected/e8bcad89-0f86-4a07-9c81-aa8cac743319-kube-api-access-htmd2\") pod \"console-f9d7485db-xq5mm\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.410251 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fqb4s"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.414255 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dtjt9"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.420080 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" event={"ID":"f9365b5c-a2b4-464c-aef6-afd71c5d3427","Type":"ContainerStarted","Data":"0a9cbdd2ccc2e6375518e7ba8ed3010b81b21e41e1dcd3e3d51391b2a370a93b"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.420112 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" event={"ID":"f9365b5c-a2b4-464c-aef6-afd71c5d3427","Type":"ContainerStarted","Data":"6b61c5ef92d84df5681d150821515793325c11f124c3b6fd0d9b7cab2585f349"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.427344 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.436029 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kcwwb" event={"ID":"8f35439c-e788-4e19-8054-6a365d28a76d","Type":"ContainerStarted","Data":"a7601a1cc179f2a80d4f3fb40106a92108708de931ba70c309d287f634133ebe"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.440402 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" event={"ID":"25f468fc-1ca1-461d-a077-2150d42bcafe","Type":"ContainerStarted","Data":"8f77b86d272f188b9d8fdb4b1716b3db96c1dff09dae49900e50c44038491f77"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.440455 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" event={"ID":"25f468fc-1ca1-461d-a077-2150d42bcafe","Type":"ContainerStarted","Data":"b43d52a2fa46cbff659b35fbedb337404e8816d5c2092a0fc9583620230efff6"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.446563 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:06 crc kubenswrapper[4745]: E1010 13:20:06.447239 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:06.947165112 +0000 UTC m=+120.844821885 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.447328 4745 generic.go:334] "Generic (PLEG): container finished" podID="f354b96e-de50-4ad1-8a82-60d570b1c6b5" containerID="cfc5d4084bcad337fca6fafecb0e215948e1f841805e1b81841dabba8514791f" exitCode=0 Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.447508 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" event={"ID":"f354b96e-de50-4ad1-8a82-60d570b1c6b5","Type":"ContainerDied","Data":"cfc5d4084bcad337fca6fafecb0e215948e1f841805e1b81841dabba8514791f"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.447551 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" event={"ID":"f354b96e-de50-4ad1-8a82-60d570b1c6b5","Type":"ContainerStarted","Data":"f0984b097013ad0f8bf99aa835eb997b9408467cbfd0c2a3989d4b79a00f2ae4"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.478700 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-kth6t" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.479860 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hqxgq" event={"ID":"0b152ecd-75c1-4677-8f8a-c27ff65d76d9","Type":"ContainerStarted","Data":"1892c577f6e9ed450fee314e6dd2fd1e944b091c03cd80febcd470bd19281f8b"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.484440 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-zdbqc"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.497441 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.502952 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.505296 4745 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-nwvhl container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.505367 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" podUID="f9365b5c-a2b4-464c-aef6-afd71c5d3427" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.505470 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-72w5j" event={"ID":"a288a2bb-c87a-464b-b2fe-d2cab9a5df70","Type":"ContainerStarted","Data":"2c6c9571b7ebb83e8e659279d0e09746fde4c2555364b406fc83a48e162c629c"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.509394 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff" event={"ID":"f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75","Type":"ContainerStarted","Data":"cfaf8fcf4fdc8c7ca43b49f690c749dc92100b2c9a0bddead70bb0d442a65de2"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.512190 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" event={"ID":"88252921-5e6b-4e51-95e8-59dd9dd891b6","Type":"ContainerStarted","Data":"01408828eaf8c970ec1daad9d93dd20b421b08d57796dca75cc2887918888afb"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.521777 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dfjd5" event={"ID":"7ddc713d-4db8-4e77-acd2-0936a8e7bcb1","Type":"ContainerStarted","Data":"b19c4a6323d5713d20ea665eca20a29b6e1a3191c9dcb01be347a5d07019abe7"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.521807 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dfjd5" event={"ID":"7ddc713d-4db8-4e77-acd2-0936a8e7bcb1","Type":"ContainerStarted","Data":"3134bdea9b11f6a344ef3b03c369c97378fd748f887fbb57f570eeb79dc206dd"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.522547 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-npfv4" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.537330 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6gqzf" event={"ID":"e0c14e43-fd09-45b2-9820-abb38a0e7761","Type":"ContainerStarted","Data":"d2a35d8053d7ecf1d8a893fd8bd0f4824a32a1eb04e89a9f240259da78c16b9f"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.537364 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6gqzf" event={"ID":"e0c14e43-fd09-45b2-9820-abb38a0e7761","Type":"ContainerStarted","Data":"cf22d635783f2d1a9bf5c42010b32118908e88b9f929d463311e65c7dcde0915"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.539782 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" event={"ID":"800c6118-3c93-4dd4-b7e4-17f23b82de1d","Type":"ContainerStarted","Data":"86972d7dd4a94619a93e1daf8a9b55df8f64c81212c10d5ba02dfd8cbb281ee3"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.543407 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" event={"ID":"60290e46-da36-4e07-a1db-edc753eb0788","Type":"ContainerStarted","Data":"607f8c419a3e6b112eb68700a4e51d8a8ed81b0f6a8a885622868cff2eadb3bf"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.543450 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" event={"ID":"60290e46-da36-4e07-a1db-edc753eb0788","Type":"ContainerStarted","Data":"2240e0012d25c1bea60bf09a57c92e5d9537f98ad63628788189ac1f4138d290"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.547760 4745 generic.go:334] "Generic (PLEG): container finished" podID="39f07a80-8096-439f-a3b5-b4dca2842b90" containerID="870b76b25c8fa3048af09c485d4c75841da833e833b224b4fc7329055f704038" exitCode=0 Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.547844 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-76fgg" event={"ID":"39f07a80-8096-439f-a3b5-b4dca2842b90","Type":"ContainerDied","Data":"870b76b25c8fa3048af09c485d4c75841da833e833b224b4fc7329055f704038"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.547875 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-76fgg" event={"ID":"39f07a80-8096-439f-a3b5-b4dca2842b90","Type":"ContainerStarted","Data":"db9b423ad8eb9ced3db08639d043039eecb7d178c07141905d5aac0fb23b183c"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.547902 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:06 crc kubenswrapper[4745]: E1010 13:20:06.549210 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:07.049195651 +0000 UTC m=+120.946852414 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.553401 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5ftts" event={"ID":"f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9","Type":"ContainerStarted","Data":"3a88d1035568306bf8bada4727ae01f83e4a965f4034ce6f87ce4d0503b658fd"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.553434 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5ftts" event={"ID":"f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9","Type":"ContainerStarted","Data":"934e257b4e1aa02ea1104af93f35bb8bf27484c3b5df6c5d9a4b1a53ad12688e"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.558857 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.573113 4745 patch_prober.go:28] interesting pod/router-default-5444994796-5ftts container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 13:20:06 crc kubenswrapper[4745]: [-]has-synced failed: reason withheld Oct 10 13:20:06 crc kubenswrapper[4745]: [+]process-running ok Oct 10 13:20:06 crc kubenswrapper[4745]: healthz check failed Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.573158 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5ftts" podUID="f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.580601 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j6gkw" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.618567 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-7v4cv" event={"ID":"420a8490-c5b6-4d9b-a716-efc526b5f4a5","Type":"ContainerStarted","Data":"b820961de8872994490ba261d5ddabafb1d76d8152dab47ec8c07a150f66598b"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.618610 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-7v4cv" event={"ID":"420a8490-c5b6-4d9b-a716-efc526b5f4a5","Type":"ContainerStarted","Data":"46759de7c122c9cc5e7d667b95d616699cc87c993b69ff6b4ee2e5c67f03ac13"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.619128 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-7v4cv" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.619872 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7jjxn" event={"ID":"247274a0-c2f0-4150-ad4f-cf39c2003a60","Type":"ContainerStarted","Data":"841ea1c0ebea5b6e5d5571353fd341d6b8289250d463bc6d81743585e5abd693"} Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.624100 4745 patch_prober.go:28] interesting pod/console-operator-58897d9998-7v4cv container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.624132 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-7v4cv" podUID="420a8490-c5b6-4d9b-a716-efc526b5f4a5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.649777 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:06 crc kubenswrapper[4745]: E1010 13:20:06.652705 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:07.15268397 +0000 UTC m=+121.050340733 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.715438 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.752917 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:06 crc kubenswrapper[4745]: E1010 13:20:06.763849 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:07.263813324 +0000 UTC m=+121.161470087 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.769410 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-gffs7"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.796164 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-rwdj4"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.809277 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.848471 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ncgfh"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.849060 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jr8mx"] Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.854090 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:06 crc kubenswrapper[4745]: E1010 13:20:06.854543 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:07.354524484 +0000 UTC m=+121.252181247 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.956528 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:06 crc kubenswrapper[4745]: E1010 13:20:06.959232 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:07.459212687 +0000 UTC m=+121.356869450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:06 crc kubenswrapper[4745]: I1010 13:20:06.970374 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-kth6t"] Oct 10 13:20:06 crc kubenswrapper[4745]: W1010 13:20:06.982802 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70186f6c_521c_49d0_a772_025dfb2deb76.slice/crio-20727877b5e7aeef0c2b408b3e8cd14b21aeae8a57827e938b8e493f5ecaac44 WatchSource:0}: Error finding container 20727877b5e7aeef0c2b408b3e8cd14b21aeae8a57827e938b8e493f5ecaac44: Status 404 returned error can't find the container with id 20727877b5e7aeef0c2b408b3e8cd14b21aeae8a57827e938b8e493f5ecaac44 Oct 10 13:20:06 crc kubenswrapper[4745]: W1010 13:20:06.988571 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03110c82_b25a_4ed8_a9ae_b59bd8181baf.slice/crio-d7eb9ae51254b62bd96c9d2659b37be20b43a6c8fee3a301f3b0f05cf6d8f686 WatchSource:0}: Error finding container d7eb9ae51254b62bd96c9d2659b37be20b43a6c8fee3a301f3b0f05cf6d8f686: Status 404 returned error can't find the container with id d7eb9ae51254b62bd96c9d2659b37be20b43a6c8fee3a301f3b0f05cf6d8f686 Oct 10 13:20:06 crc kubenswrapper[4745]: W1010 13:20:06.989230 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65704ac7_fe35_4513_85b3_d512314162e8.slice/crio-babdc68fac434a19d2e147e410f6174d0a34e73b4bc3315c36119b06971113e2 WatchSource:0}: Error finding container babdc68fac434a19d2e147e410f6174d0a34e73b4bc3315c36119b06971113e2: Status 404 returned error can't find the container with id babdc68fac434a19d2e147e410f6174d0a34e73b4bc3315c36119b06971113e2 Oct 10 13:20:06 crc kubenswrapper[4745]: W1010 13:20:06.993019 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c8c7598_c1b9_4201_9f79_21c0c312f26a.slice/crio-a2315a196382c0da96a207fffe175e5191bdebca9c3d7f5672af5c1301078e60 WatchSource:0}: Error finding container a2315a196382c0da96a207fffe175e5191bdebca9c3d7f5672af5c1301078e60: Status 404 returned error can't find the container with id a2315a196382c0da96a207fffe175e5191bdebca9c3d7f5672af5c1301078e60 Oct 10 13:20:06 crc kubenswrapper[4745]: W1010 13:20:06.994172 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod089f446f_66e5_4e2b_99cd_9d72f05c0990.slice/crio-32dd2149654e5a672750be4f76e3771d696a81f9550a6185f2a56f6947802c24 WatchSource:0}: Error finding container 32dd2149654e5a672750be4f76e3771d696a81f9550a6185f2a56f6947802c24: Status 404 returned error can't find the container with id 32dd2149654e5a672750be4f76e3771d696a81f9550a6185f2a56f6947802c24 Oct 10 13:20:07 crc kubenswrapper[4745]: W1010 13:20:07.036898 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bca2f23_df4e_4144_8253_9ad9b08df710.slice/crio-d93e78b969fb08d3afe391cfe583dab3d65504d78707a78feebb6d66287841d5 WatchSource:0}: Error finding container d93e78b969fb08d3afe391cfe583dab3d65504d78707a78feebb6d66287841d5: Status 404 returned error can't find the container with id d93e78b969fb08d3afe391cfe583dab3d65504d78707a78feebb6d66287841d5 Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.057707 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:07 crc kubenswrapper[4745]: E1010 13:20:07.058407 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:07.558378504 +0000 UTC m=+121.456035277 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.058640 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:07 crc kubenswrapper[4745]: E1010 13:20:07.059001 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:07.558993612 +0000 UTC m=+121.456650375 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.066206 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8m4nd"] Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.160439 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:07 crc kubenswrapper[4745]: E1010 13:20:07.160883 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:07.660867885 +0000 UTC m=+121.558524648 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.262309 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:07 crc kubenswrapper[4745]: E1010 13:20:07.263108 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:07.763079149 +0000 UTC m=+121.660735912 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.314691 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lwl2w"] Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.354393 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-7v4cv" podStartSLOduration=100.354364055 podStartE2EDuration="1m40.354364055s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:07.353080719 +0000 UTC m=+121.250737492" watchObservedRunningTime="2025-10-10 13:20:07.354364055 +0000 UTC m=+121.252020818" Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.364137 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:07 crc kubenswrapper[4745]: E1010 13:20:07.364573 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:07.864558231 +0000 UTC m=+121.762214994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.467122 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:07 crc kubenswrapper[4745]: E1010 13:20:07.471766 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:07.971742224 +0000 UTC m=+121.869398987 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.569478 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:07 crc kubenswrapper[4745]: E1010 13:20:07.570036 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:08.070021857 +0000 UTC m=+121.967678620 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.573458 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" podStartSLOduration=99.573438423 podStartE2EDuration="1m39.573438423s" podCreationTimestamp="2025-10-10 13:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:07.538843121 +0000 UTC m=+121.436499894" watchObservedRunningTime="2025-10-10 13:20:07.573438423 +0000 UTC m=+121.471095176" Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.610813 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-npfv4"] Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.648454 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lwl2w" event={"ID":"08808841-2872-4cab-8ccd-208e1d29e3b0","Type":"ContainerStarted","Data":"d29e3355f55e359774bb77ec435ede4a2c9d2474c8890ae7557975a4cdad0942"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.656796 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" event={"ID":"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3","Type":"ContainerStarted","Data":"dc9a1d275d3a25763aeb84005a5a2a0687c95c3ab2289f74082126002bea01f7"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.670208 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj" event={"ID":"70186f6c-521c-49d0-a772-025dfb2deb76","Type":"ContainerStarted","Data":"20727877b5e7aeef0c2b408b3e8cd14b21aeae8a57827e938b8e493f5ecaac44"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.674340 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:07 crc kubenswrapper[4745]: E1010 13:20:07.674788 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:08.174777192 +0000 UTC m=+122.072433955 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.682713 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ncgfh" event={"ID":"c0a536ca-a2c3-4cbf-9a0d-54fb68f80510","Type":"ContainerStarted","Data":"bd5898713543b78267e6c64c2149377658fead4056290f3e0071d600aa934310"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.689274 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-zdbqc" event={"ID":"469f0496-8d18-43a5-b25f-a9450f92fac7","Type":"ContainerStarted","Data":"1826c67e127b00032c0e39cdd5eab248677de36fdbbb728afb5502ba13cac8c1"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.689474 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-zdbqc" event={"ID":"469f0496-8d18-43a5-b25f-a9450f92fac7","Type":"ContainerStarted","Data":"4e362ed722bdb1072122b080b5c04665451ccb5b868b6c3dd66828ce74fbe118"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.718599 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kcwwb" event={"ID":"8f35439c-e788-4e19-8054-6a365d28a76d","Type":"ContainerStarted","Data":"f9ea33ce3e57073beba5f1479632c347f17d06f15fb13364401d142e78780464"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.732650 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jr8mx" event={"ID":"089f446f-66e5-4e2b-99cd-9d72f05c0990","Type":"ContainerStarted","Data":"32dd2149654e5a672750be4f76e3771d696a81f9550a6185f2a56f6947802c24"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.760189 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-5ftts" podStartSLOduration=100.760171723 podStartE2EDuration="1m40.760171723s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:07.758468795 +0000 UTC m=+121.656125568" watchObservedRunningTime="2025-10-10 13:20:07.760171723 +0000 UTC m=+121.657828486" Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.760299 4745 patch_prober.go:28] interesting pod/router-default-5444994796-5ftts container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 13:20:07 crc kubenswrapper[4745]: [-]has-synced failed: reason withheld Oct 10 13:20:07 crc kubenswrapper[4745]: [+]process-running ok Oct 10 13:20:07 crc kubenswrapper[4745]: healthz check failed Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.760351 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5ftts" podUID="f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.776198 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:07 crc kubenswrapper[4745]: E1010 13:20:07.776629 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:08.276606035 +0000 UTC m=+122.174262798 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.805698 4745 generic.go:334] "Generic (PLEG): container finished" podID="a288a2bb-c87a-464b-b2fe-d2cab9a5df70" containerID="9a71daefe6a6a9ba7eca20a86d8d8c66daaa0acf61532041b617321b097c2dc2" exitCode=0 Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.805836 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-72w5j" event={"ID":"a288a2bb-c87a-464b-b2fe-d2cab9a5df70","Type":"ContainerDied","Data":"9a71daefe6a6a9ba7eca20a86d8d8c66daaa0acf61532041b617321b097c2dc2"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.820307 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff" event={"ID":"f2f4d74f-6bcf-4bb0-871f-b0e8e5e65d75","Type":"ContainerStarted","Data":"5149edb21a818eda29b8c6e475e9043dcf4a07e9947c3097d6877c37fb637b99"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.838766 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6gqzf" podStartSLOduration=100.838745391 podStartE2EDuration="1m40.838745391s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:07.838025051 +0000 UTC m=+121.735681814" watchObservedRunningTime="2025-10-10 13:20:07.838745391 +0000 UTC m=+121.736402154" Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.860145 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" event={"ID":"4c8c7598-c1b9-4201-9f79-21c0c312f26a","Type":"ContainerStarted","Data":"a2315a196382c0da96a207fffe175e5191bdebca9c3d7f5672af5c1301078e60"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.862606 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" event={"ID":"f9346bb5-09f3-4539-b678-198f2e8da9a4","Type":"ContainerStarted","Data":"21412557a77541b2c037bdd3a38a29f2e9f04e8673d3738f89815ea098f1d1b5"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.864384 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-pdt5f" event={"ID":"e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef","Type":"ContainerStarted","Data":"9f2d5fb43cb31b737c1fcb1d0d3a76dbafca17b1155b5b61da9c7820dbf9f8c2"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.866793 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" event={"ID":"800c6118-3c93-4dd4-b7e4-17f23b82de1d","Type":"ContainerStarted","Data":"d18d0e4826d4258a3004b21923525c803a3194811eeb9d7fb6231fdb05267bcf"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.870873 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" event={"ID":"60290e46-da36-4e07-a1db-edc753eb0788","Type":"ContainerStarted","Data":"8469561ecf9b4c254dc487da0ca519d99cbcfe64adbebb9a4adb0da6323235c7"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.879876 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:07 crc kubenswrapper[4745]: E1010 13:20:07.882587 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:08.382560883 +0000 UTC m=+122.280217646 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.889589 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fqb4s" event={"ID":"839da6a0-320a-45d1-8a6c-be916ee27421","Type":"ContainerStarted","Data":"48950f39b119d3fc2b013dca168d9db51cde5863bc0b31c84ab08a02604ff1d9"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.889641 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fqb4s" event={"ID":"839da6a0-320a-45d1-8a6c-be916ee27421","Type":"ContainerStarted","Data":"7db11ebecb644e32ce6acd3447605df7fe214bfb70d1cb033e8a687fd745d4e3"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.897430 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" event={"ID":"0088430d-3a96-46ea-a149-7c59eb24d706","Type":"ContainerStarted","Data":"053f3395c80cb98947103d8af9aa61e1ced574f401c42d7560e7d02647094611"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.897476 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" event={"ID":"0088430d-3a96-46ea-a149-7c59eb24d706","Type":"ContainerStarted","Data":"0242f4f578af6c17968a8008de59ed305086c659e8032c1930a698c69e1bf77c"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.897722 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.903820 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dtjt9" event={"ID":"d159f38f-c259-44cf-a5f9-8c4911d7864c","Type":"ContainerStarted","Data":"7f7574865e60d3ff369491273fef0ccb59d9612f760af93f2f0a230dc60b0b02"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.906207 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" event={"ID":"03110c82-b25a-4ed8-a9ae-b59bd8181baf","Type":"ContainerStarted","Data":"d7eb9ae51254b62bd96c9d2659b37be20b43a6c8fee3a301f3b0f05cf6d8f686"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.908674 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" event={"ID":"c157d5c1-d9e5-4b0d-94b7-432bb85ffa10","Type":"ContainerStarted","Data":"5a5660c1c5c6957d74f4e40e76c286f8e920b2aaa52161a10bc8c11887f52c77"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.910225 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gffs7" event={"ID":"65704ac7-fe35-4513-85b3-d512314162e8","Type":"ContainerStarted","Data":"babdc68fac434a19d2e147e410f6174d0a34e73b4bc3315c36119b06971113e2"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.912851 4745 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-9q9xb container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.912897 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" podUID="0088430d-3a96-46ea-a149-7c59eb24d706" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.919632 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-kth6t" event={"ID":"1bca2f23-df4e-4144-8253-9ad9b08df710","Type":"ContainerStarted","Data":"d93e78b969fb08d3afe391cfe583dab3d65504d78707a78feebb6d66287841d5"} Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.925330 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.985559 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:07 crc kubenswrapper[4745]: E1010 13:20:07.985711 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:08.485684442 +0000 UTC m=+122.383341205 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:07 crc kubenswrapper[4745]: E1010 13:20:07.989238 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:08.489138389 +0000 UTC m=+122.386795352 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:07 crc kubenswrapper[4745]: I1010 13:20:07.993470 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.093467 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-7v4cv" Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.102253 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:08 crc kubenswrapper[4745]: E1010 13:20:08.103060 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:08.603033911 +0000 UTC m=+122.500690674 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.155718 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dfjd5" podStartSLOduration=101.155444224 podStartE2EDuration="1m41.155444224s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:08.145368371 +0000 UTC m=+122.043025134" watchObservedRunningTime="2025-10-10 13:20:08.155444224 +0000 UTC m=+122.053100977" Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.204191 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:08 crc kubenswrapper[4745]: E1010 13:20:08.204462 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:08.704449632 +0000 UTC m=+122.602106395 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.254874 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vrllq"] Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.285568 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-kgpj4"] Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.304827 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:08 crc kubenswrapper[4745]: E1010 13:20:08.304969 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:08.804943227 +0000 UTC m=+122.702599990 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.305095 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:08 crc kubenswrapper[4745]: E1010 13:20:08.305354 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:08.805342658 +0000 UTC m=+122.702999421 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.386509 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fqb4s" podStartSLOduration=101.386465518 podStartE2EDuration="1m41.386465518s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:08.38403238 +0000 UTC m=+122.281689143" watchObservedRunningTime="2025-10-10 13:20:08.386465518 +0000 UTC m=+122.284122291" Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.405868 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:08 crc kubenswrapper[4745]: E1010 13:20:08.406177 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:08.906140202 +0000 UTC m=+122.803796965 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.406339 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:08 crc kubenswrapper[4745]: E1010 13:20:08.406717 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:08.906707247 +0000 UTC m=+122.804364020 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.420841 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pg69q" podStartSLOduration=101.420826384 podStartE2EDuration="1m41.420826384s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:08.419008003 +0000 UTC m=+122.316664776" watchObservedRunningTime="2025-10-10 13:20:08.420826384 +0000 UTC m=+122.318483147" Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.448886 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j6gkw"] Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.467806 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4"] Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.474656 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w"] Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.480503 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf"] Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.498506 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dc27t"] Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.507915 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:08 crc kubenswrapper[4745]: E1010 13:20:08.509059 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:09.009039044 +0000 UTC m=+122.906695807 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.537954 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fk8ff" podStartSLOduration=101.537938547 podStartE2EDuration="1m41.537938547s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:08.535102377 +0000 UTC m=+122.432759140" watchObservedRunningTime="2025-10-10 13:20:08.537938547 +0000 UTC m=+122.435595310" Oct 10 13:20:08 crc kubenswrapper[4745]: W1010 13:20:08.552273 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod340979c3_ead3_4393_9aeb_3c6957ccdf80.slice/crio-0516581ea4a8242d424e7158b40b2b1ce41fd11e7d9b5fa18d41181a21d4df31 WatchSource:0}: Error finding container 0516581ea4a8242d424e7158b40b2b1ce41fd11e7d9b5fa18d41181a21d4df31: Status 404 returned error can't find the container with id 0516581ea4a8242d424e7158b40b2b1ce41fd11e7d9b5fa18d41181a21d4df31 Oct 10 13:20:08 crc kubenswrapper[4745]: W1010 13:20:08.552559 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cb8607d_1cbd_4b3a_9b06_f2db94952d82.slice/crio-b17eb1972d0373c465ad9fdf8d4491f220a3eda08881637d0874b1713d219818 WatchSource:0}: Error finding container b17eb1972d0373c465ad9fdf8d4491f220a3eda08881637d0874b1713d219818: Status 404 returned error can't find the container with id b17eb1972d0373c465ad9fdf8d4491f220a3eda08881637d0874b1713d219818 Oct 10 13:20:08 crc kubenswrapper[4745]: W1010 13:20:08.554428 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8a4af69_070d_48f1_8f56_1466adc9efb6.slice/crio-fe32deed460d7dcd797788d6b71e65a0e062875356acd65ebacda3da6b65e5ec WatchSource:0}: Error finding container fe32deed460d7dcd797788d6b71e65a0e062875356acd65ebacda3da6b65e5ec: Status 404 returned error can't find the container with id fe32deed460d7dcd797788d6b71e65a0e062875356acd65ebacda3da6b65e5ec Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.571096 4745 patch_prober.go:28] interesting pod/router-default-5444994796-5ftts container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 13:20:08 crc kubenswrapper[4745]: [-]has-synced failed: reason withheld Oct 10 13:20:08 crc kubenswrapper[4745]: [+]process-running ok Oct 10 13:20:08 crc kubenswrapper[4745]: healthz check failed Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.571154 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5ftts" podUID="f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 13:20:08 crc kubenswrapper[4745]: W1010 13:20:08.571718 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bff784a_c53a_4ff1_8507_e45b2eab9f1f.slice/crio-50cfadd13bc8e509aee08d0f5bfd91be36723989cb7f05ed99a76a48338a9805 WatchSource:0}: Error finding container 50cfadd13bc8e509aee08d0f5bfd91be36723989cb7f05ed99a76a48338a9805: Status 404 returned error can't find the container with id 50cfadd13bc8e509aee08d0f5bfd91be36723989cb7f05ed99a76a48338a9805 Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.587664 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" podStartSLOduration=101.587615463 podStartE2EDuration="1m41.587615463s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:08.58429079 +0000 UTC m=+122.481947553" watchObservedRunningTime="2025-10-10 13:20:08.587615463 +0000 UTC m=+122.485272226" Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.614185 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:08 crc kubenswrapper[4745]: E1010 13:20:08.614540 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:09.11452537 +0000 UTC m=+123.012182133 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.683157 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-xq5mm"] Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.715070 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:08 crc kubenswrapper[4745]: E1010 13:20:08.716426 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:09.216410954 +0000 UTC m=+123.114067717 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.817991 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:08 crc kubenswrapper[4745]: E1010 13:20:08.818472 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:09.318447052 +0000 UTC m=+123.216103815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.919911 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:08 crc kubenswrapper[4745]: E1010 13:20:08.920407 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:09.420387178 +0000 UTC m=+123.318043941 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.965328 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-76fgg" event={"ID":"39f07a80-8096-439f-a3b5-b4dca2842b90","Type":"ContainerStarted","Data":"6881c76588350b20824ca74b51a3fe41db99c286a4182e8582bb22443161a809"} Oct 10 13:20:08 crc kubenswrapper[4745]: I1010 13:20:08.966327 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-76fgg" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.023455 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.024072 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j6gkw" event={"ID":"870f50c7-a485-45ae-8911-eaa1aa314e44","Type":"ContainerStarted","Data":"57a7619a58d784bc18573a023bc9fdac3454f4a1043f5872c5090a39692bd855"} Oct 10 13:20:09 crc kubenswrapper[4745]: E1010 13:20:09.024315 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:09.524298149 +0000 UTC m=+123.421954912 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.091167 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" event={"ID":"4c8c7598-c1b9-4201-9f79-21c0c312f26a","Type":"ContainerStarted","Data":"4178cab18529594c9c15b5d31f2d138d7d767a520a40a9261051d6022176d028"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.091511 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.097439 4745 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-bmlmb container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" start-of-body= Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.097485 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" podUID="4c8c7598-c1b9-4201-9f79-21c0c312f26a" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.102303 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-kth6t" event={"ID":"1bca2f23-df4e-4144-8253-9ad9b08df710","Type":"ContainerStarted","Data":"a3df400a3b0c88af99acfd33bed58c99c92b623ec3010ed6ad0abc389dd0659e"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.113068 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" podStartSLOduration=101.113053654 podStartE2EDuration="1m41.113053654s" podCreationTimestamp="2025-10-10 13:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.112294222 +0000 UTC m=+123.009950975" watchObservedRunningTime="2025-10-10 13:20:09.113053654 +0000 UTC m=+123.010710417" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.113945 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-76fgg" podStartSLOduration=102.113938699 podStartE2EDuration="1m42.113938699s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.011768466 +0000 UTC m=+122.909425229" watchObservedRunningTime="2025-10-10 13:20:09.113938699 +0000 UTC m=+123.011595462" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.114712 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7jjxn" event={"ID":"247274a0-c2f0-4150-ad4f-cf39c2003a60","Type":"ContainerStarted","Data":"b6628b061c5856de44cd30885792f953ff91e7055abe9b3b4313d60001875b90"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.115916 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-7jjxn" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.124711 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:09 crc kubenswrapper[4745]: E1010 13:20:09.126142 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:09.626126121 +0000 UTC m=+123.523782874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.126599 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" event={"ID":"5cb8607d-1cbd-4b3a-9b06-f2db94952d82","Type":"ContainerStarted","Data":"b17eb1972d0373c465ad9fdf8d4491f220a3eda08881637d0874b1713d219818"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.126610 4745 patch_prober.go:28] interesting pod/downloads-7954f5f757-7jjxn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.126827 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7jjxn" podUID="247274a0-c2f0-4150-ad4f-cf39c2003a60" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.133000 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" event={"ID":"88252921-5e6b-4e51-95e8-59dd9dd891b6","Type":"ContainerStarted","Data":"33acc96f22e56c1a76bfc4ea40bddb4c69308dfb3e3ba3c68ad35d798739dbfc"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.144421 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-7jjxn" podStartSLOduration=102.144404355 podStartE2EDuration="1m42.144404355s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.14315778 +0000 UTC m=+123.040814543" watchObservedRunningTime="2025-10-10 13:20:09.144404355 +0000 UTC m=+123.042061118" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.148935 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf" event={"ID":"a8a4af69-070d-48f1-8f56-1466adc9efb6","Type":"ContainerStarted","Data":"fe32deed460d7dcd797788d6b71e65a0e062875356acd65ebacda3da6b65e5ec"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.152981 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xq5mm" event={"ID":"e8bcad89-0f86-4a07-9c81-aa8cac743319","Type":"ContainerStarted","Data":"89e93c8e1cadc332351e61a749998383794f241505d7ca3c02dc02a103dc2b12"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.163919 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-bgmxs" podStartSLOduration=102.163894273 podStartE2EDuration="1m42.163894273s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.162889165 +0000 UTC m=+123.060545928" watchObservedRunningTime="2025-10-10 13:20:09.163894273 +0000 UTC m=+123.061551036" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.218818 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vrllq" event={"ID":"97baf9cf-d52a-421e-b6b5-5996d71adf41","Type":"ContainerStarted","Data":"7e78571db7c2e95444af7557fb65430f2b23f8cce230364c85947e3e2ba6e4ae"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.218870 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vrllq" event={"ID":"97baf9cf-d52a-421e-b6b5-5996d71adf41","Type":"ContainerStarted","Data":"d53567f658b2fe09f7e1939114576045e3019e4a38eeffd32cf598aa0e0ebdbf"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.228866 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:09 crc kubenswrapper[4745]: E1010 13:20:09.229911 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:09.729897268 +0000 UTC m=+123.627554031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.241211 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" event={"ID":"800c6118-3c93-4dd4-b7e4-17f23b82de1d","Type":"ContainerStarted","Data":"c8072f64a663b304504e837cfeeb9ba311ece9fc2367d71f51ab07d0cbe0a2b4"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.251996 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vrllq" podStartSLOduration=101.251970679 podStartE2EDuration="1m41.251970679s" podCreationTimestamp="2025-10-10 13:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.248956484 +0000 UTC m=+123.146613247" watchObservedRunningTime="2025-10-10 13:20:09.251970679 +0000 UTC m=+123.149627442" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.266427 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ncgfh" event={"ID":"c0a536ca-a2c3-4cbf-9a0d-54fb68f80510","Type":"ContainerStarted","Data":"37122998a360f929c4c29039e49b9e18f114f195fa1c3d0dce1dc312773f8f50"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.283303 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-s8wbg" podStartSLOduration=102.283286569 podStartE2EDuration="1m42.283286569s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.281256702 +0000 UTC m=+123.178913485" watchObservedRunningTime="2025-10-10 13:20:09.283286569 +0000 UTC m=+123.180943332" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.299641 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" event={"ID":"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3","Type":"ContainerStarted","Data":"539d549721a124f489450d0d93f931208c97452ef2c7fad3f880aee6231ba8da"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.300523 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.311615 4745 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-98dn4 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.25:6443/healthz\": dial tcp 10.217.0.25:6443: connect: connection refused" start-of-body= Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.311701 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" podUID="11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.25:6443/healthz\": dial tcp 10.217.0.25:6443: connect: connection refused" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.331034 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:09 crc kubenswrapper[4745]: E1010 13:20:09.331280 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:09.831248177 +0000 UTC m=+123.728904940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.331819 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-npfv4" event={"ID":"cfa4b474-6bd6-4cc3-910e-d1199af5f7c6","Type":"ContainerStarted","Data":"1267733697bfc532749eb25ec20bc4f7039379e5de1014ddc4437b830137a462"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.331859 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-npfv4" event={"ID":"cfa4b474-6bd6-4cc3-910e-d1199af5f7c6","Type":"ContainerStarted","Data":"a148078fd85e15fa905ec303f811a06a6c1ae2ce3abd47cecb0f827e986d6334"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.332401 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:09 crc kubenswrapper[4745]: E1010 13:20:09.333563 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:09.833553192 +0000 UTC m=+123.731209955 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.346956 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4" event={"ID":"340979c3-ead3-4393-9aeb-3c6957ccdf80","Type":"ContainerStarted","Data":"0516581ea4a8242d424e7158b40b2b1ce41fd11e7d9b5fa18d41181a21d4df31"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.347864 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.372486 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" podStartSLOduration=102.372469376 podStartE2EDuration="1m42.372469376s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.331935127 +0000 UTC m=+123.229591890" watchObservedRunningTime="2025-10-10 13:20:09.372469376 +0000 UTC m=+123.270126139" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.374220 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-npfv4" podStartSLOduration=102.374214015 podStartE2EDuration="1m42.374214015s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.364042649 +0000 UTC m=+123.261699412" watchObservedRunningTime="2025-10-10 13:20:09.374214015 +0000 UTC m=+123.271870778" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.382710 4745 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-skrn4 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.382823 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4" podUID="340979c3-ead3-4393-9aeb-3c6957ccdf80" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.396277 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kcwwb" event={"ID":"8f35439c-e788-4e19-8054-6a365d28a76d","Type":"ContainerStarted","Data":"12b920b941407ea2d62f76177ab71f19c3a7d174ddb2c4a833050e7ee2671cff"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.426132 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-pdt5f" event={"ID":"e1c9e5e0-0383-4bf2-9fc0-bff5bca8d1ef","Type":"ContainerStarted","Data":"dfecf6232054041e0b0bfbd53d0669ed0f4c912061174cfbbfb040dea841f49a"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.433245 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:09 crc kubenswrapper[4745]: E1010 13:20:09.434307 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:09.934292784 +0000 UTC m=+123.831949547 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.436497 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj" event={"ID":"70186f6c-521c-49d0-a772-025dfb2deb76","Type":"ContainerStarted","Data":"d49b5d7b046af9b8e8b4dce74649124f225df7433d59ea728af4feda7d47c70e"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.437426 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.444581 4745 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-tnbzj container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.444645 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj" podUID="70186f6c-521c-49d0-a772-025dfb2deb76" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.452275 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gffs7" event={"ID":"65704ac7-fe35-4513-85b3-d512314162e8","Type":"ContainerStarted","Data":"ffe1aa6eda0fa342ba8b2fedfe74dd5cb0d923908540268b230c37f80233c156"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.468042 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lwl2w" event={"ID":"08808841-2872-4cab-8ccd-208e1d29e3b0","Type":"ContainerStarted","Data":"16c538f5e577f1b7b719588f11495dd8547b9aec7a376c1307566d8fea9b4d2e"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.473282 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" event={"ID":"7bff784a-c53a-4ff1-8507-e45b2eab9f1f","Type":"ContainerStarted","Data":"50cfadd13bc8e509aee08d0f5bfd91be36723989cb7f05ed99a76a48338a9805"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.474246 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.475475 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" event={"ID":"25f468fc-1ca1-461d-a077-2150d42bcafe","Type":"ContainerStarted","Data":"a55c39da360d62db68de45b408cb177b09b927ba4da76de9690e2460eb58abd9"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.476810 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jr8mx" event={"ID":"089f446f-66e5-4e2b-99cd-9d72f05c0990","Type":"ContainerStarted","Data":"9ca521c575814a192c32ee3029cfc81f250cc8a8c8f51c6c8d4d78980078b33c"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.478094 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-kgpj4" event={"ID":"83aaa6eb-34bb-4755-a8af-bd5925aef46c","Type":"ContainerStarted","Data":"876dd622d47a111e0cf1f6b662a77839904c48b6f1930c913d2d270f316ac6c9"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.478123 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-kgpj4" event={"ID":"83aaa6eb-34bb-4755-a8af-bd5925aef46c","Type":"ContainerStarted","Data":"1854b7e7690913525c96d5732513f5e44b94a1abf2f9a061d6e8293d0cbe1466"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.491200 4745 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dc27t container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.491257 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" podUID="7bff784a-c53a-4ff1-8507-e45b2eab9f1f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.493357 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hqxgq" event={"ID":"0b152ecd-75c1-4677-8f8a-c27ff65d76d9","Type":"ContainerStarted","Data":"8ce72bbb051537b3f69c5d6f4be91c3f17c02c6c37b6ccf50d8e28b2c583cd9b"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.493395 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hqxgq" event={"ID":"0b152ecd-75c1-4677-8f8a-c27ff65d76d9","Type":"ContainerStarted","Data":"aa53270018f5e0d4d21d5ce3340576917f175d182db4ca2bf63bc6434083abfd"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.495035 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" event={"ID":"f9346bb5-09f3-4539-b678-198f2e8da9a4","Type":"ContainerStarted","Data":"a918c722293efe8e9da57172d0314ba309c534f2061a5a083acfced5c3b31d11"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.510396 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dtjt9" event={"ID":"d159f38f-c259-44cf-a5f9-8c4911d7864c","Type":"ContainerStarted","Data":"c601ce4af78c992cc0c1b70d48d8a2322e44ea532fe9471a96d316abd3b5b436"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.513545 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" event={"ID":"f354b96e-de50-4ad1-8a82-60d570b1c6b5","Type":"ContainerStarted","Data":"c05ba8d166dc37ceed15da00ee2393734337700290759393f9a26dbe0ca4cdf9"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.544490 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4" podStartSLOduration=101.544465401 podStartE2EDuration="1m41.544465401s" podCreationTimestamp="2025-10-10 13:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.407178652 +0000 UTC m=+123.304835415" watchObservedRunningTime="2025-10-10 13:20:09.544465401 +0000 UTC m=+123.442122164" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.544947 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kcwwb" podStartSLOduration=102.544941485 podStartE2EDuration="1m42.544941485s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.525681423 +0000 UTC m=+123.423338196" watchObservedRunningTime="2025-10-10 13:20:09.544941485 +0000 UTC m=+123.442598258" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.547672 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.556705 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-72w5j" event={"ID":"a288a2bb-c87a-464b-b2fe-d2cab9a5df70","Type":"ContainerStarted","Data":"0030925bd1d68c758ad3b8236f2eb3be30f6859654c6be54b4fc3ed592e5d659"} Oct 10 13:20:09 crc kubenswrapper[4745]: E1010 13:20:09.565422 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:10.065375309 +0000 UTC m=+123.963032072 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.576236 4745 patch_prober.go:28] interesting pod/router-default-5444994796-5ftts container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 13:20:09 crc kubenswrapper[4745]: [-]has-synced failed: reason withheld Oct 10 13:20:09 crc kubenswrapper[4745]: [+]process-running ok Oct 10 13:20:09 crc kubenswrapper[4745]: healthz check failed Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.576714 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5ftts" podUID="f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.600486 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" event={"ID":"c157d5c1-d9e5-4b0d-94b7-432bb85ffa10","Type":"ContainerStarted","Data":"97b13fc161d4c0c7c1f16865c562a882db208bad9bc2d218f599de4e44c51458"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.600525 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" event={"ID":"c157d5c1-d9e5-4b0d-94b7-432bb85ffa10","Type":"ContainerStarted","Data":"e1b8a42b17c474da5fed93d950bd8ded415fd94277e1120d7527c6386cc81113"} Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.623866 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.632065 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj" podStartSLOduration=101.632031193 podStartE2EDuration="1m41.632031193s" podCreationTimestamp="2025-10-10 13:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.592915343 +0000 UTC m=+123.490572106" watchObservedRunningTime="2025-10-10 13:20:09.632031193 +0000 UTC m=+123.529687956" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.633337 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-8m4nd" podStartSLOduration=102.633329059 podStartE2EDuration="1m42.633329059s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.633054621 +0000 UTC m=+123.530711384" watchObservedRunningTime="2025-10-10 13:20:09.633329059 +0000 UTC m=+123.530985822" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.657525 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.659662 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jr8mx" podStartSLOduration=102.659638569 podStartE2EDuration="1m42.659638569s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.659531886 +0000 UTC m=+123.557188649" watchObservedRunningTime="2025-10-10 13:20:09.659638569 +0000 UTC m=+123.557295362" Oct 10 13:20:09 crc kubenswrapper[4745]: E1010 13:20:09.661207 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:10.161187522 +0000 UTC m=+124.058844285 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.661477 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:09 crc kubenswrapper[4745]: E1010 13:20:09.662192 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:10.16218393 +0000 UTC m=+124.059840693 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.701375 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvnrp" podStartSLOduration=102.701352631 podStartE2EDuration="1m42.701352631s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.699346855 +0000 UTC m=+123.597003618" watchObservedRunningTime="2025-10-10 13:20:09.701352631 +0000 UTC m=+123.599009404" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.735243 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-kgpj4" podStartSLOduration=101.735225884 podStartE2EDuration="1m41.735225884s" podCreationTimestamp="2025-10-10 13:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.731433557 +0000 UTC m=+123.629090320" watchObservedRunningTime="2025-10-10 13:20:09.735225884 +0000 UTC m=+123.632882647" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.763330 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:09 crc kubenswrapper[4745]: E1010 13:20:09.764599 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:10.264579409 +0000 UTC m=+124.162236182 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.825810 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dtjt9" podStartSLOduration=102.825642484 podStartE2EDuration="1m42.825642484s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.786293308 +0000 UTC m=+123.683950071" watchObservedRunningTime="2025-10-10 13:20:09.825642484 +0000 UTC m=+123.723299247" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.828345 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" podStartSLOduration=101.82832289 podStartE2EDuration="1m41.82832289s" podCreationTimestamp="2025-10-10 13:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.825209262 +0000 UTC m=+123.722866015" watchObservedRunningTime="2025-10-10 13:20:09.82832289 +0000 UTC m=+123.725979653" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.866993 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:09 crc kubenswrapper[4745]: E1010 13:20:09.867495 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:10.36747808 +0000 UTC m=+124.265134843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.892615 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" podStartSLOduration=101.892591956 podStartE2EDuration="1m41.892591956s" podCreationTimestamp="2025-10-10 13:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.891438614 +0000 UTC m=+123.789095377" watchObservedRunningTime="2025-10-10 13:20:09.892591956 +0000 UTC m=+123.790248719" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.967720 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:09 crc kubenswrapper[4745]: E1010 13:20:09.968202 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:10.468187481 +0000 UTC m=+124.365844244 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.979518 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hqxgq" podStartSLOduration=102.979487769 podStartE2EDuration="1m42.979487769s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.942325684 +0000 UTC m=+123.839982437" watchObservedRunningTime="2025-10-10 13:20:09.979487769 +0000 UTC m=+123.877144532" Oct 10 13:20:09 crc kubenswrapper[4745]: I1010 13:20:09.981525 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lwl2w" podStartSLOduration=102.981518916 podStartE2EDuration="1m42.981518916s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:09.980066085 +0000 UTC m=+123.877722848" watchObservedRunningTime="2025-10-10 13:20:09.981518916 +0000 UTC m=+123.879175679" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.009862 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-pdt5f" podStartSLOduration=7.009845612 podStartE2EDuration="7.009845612s" podCreationTimestamp="2025-10-10 13:20:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:10.000850179 +0000 UTC m=+123.898506952" watchObservedRunningTime="2025-10-10 13:20:10.009845612 +0000 UTC m=+123.907502375" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.028016 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-zdbqc" podStartSLOduration=7.027992332 podStartE2EDuration="7.027992332s" podCreationTimestamp="2025-10-10 13:20:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:10.026516981 +0000 UTC m=+123.924173754" watchObservedRunningTime="2025-10-10 13:20:10.027992332 +0000 UTC m=+123.925649095" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.069745 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:10 crc kubenswrapper[4745]: E1010 13:20:10.070273 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:10.570258561 +0000 UTC m=+124.467915324 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.171352 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:10 crc kubenswrapper[4745]: E1010 13:20:10.171637 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:10.6716225 +0000 UTC m=+124.569279263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.272446 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:10 crc kubenswrapper[4745]: E1010 13:20:10.272768 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:10.772754433 +0000 UTC m=+124.670411196 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.354002 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.354063 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.373175 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:10 crc kubenswrapper[4745]: E1010 13:20:10.373514 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:10.873424043 +0000 UTC m=+124.771080806 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.373692 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:10 crc kubenswrapper[4745]: E1010 13:20:10.374031 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:10.87402287 +0000 UTC m=+124.771679633 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.475030 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:10 crc kubenswrapper[4745]: E1010 13:20:10.475215 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:10.975183593 +0000 UTC m=+124.872840356 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.475417 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:10 crc kubenswrapper[4745]: E1010 13:20:10.475761 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:10.975749609 +0000 UTC m=+124.873406382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.567502 4745 patch_prober.go:28] interesting pod/router-default-5444994796-5ftts container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 13:20:10 crc kubenswrapper[4745]: [-]has-synced failed: reason withheld Oct 10 13:20:10 crc kubenswrapper[4745]: [+]process-running ok Oct 10 13:20:10 crc kubenswrapper[4745]: healthz check failed Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.567575 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5ftts" podUID="f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.576117 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:10 crc kubenswrapper[4745]: E1010 13:20:10.576334 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:11.076291896 +0000 UTC m=+124.973948659 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.576449 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:10 crc kubenswrapper[4745]: E1010 13:20:10.577211 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:11.077184941 +0000 UTC m=+124.974841704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.608060 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4" event={"ID":"340979c3-ead3-4393-9aeb-3c6957ccdf80","Type":"ContainerStarted","Data":"20eb370859208fc6d1ad212cf9a2e10c305976c0491cff2247eaffed6c0e2780"} Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.609907 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j6gkw" event={"ID":"870f50c7-a485-45ae-8911-eaa1aa314e44","Type":"ContainerStarted","Data":"bd2ad828edc6ddba1e81a9cefd224ecb174a0f899a3828578db6c063bd7b567a"} Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.609949 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j6gkw" event={"ID":"870f50c7-a485-45ae-8911-eaa1aa314e44","Type":"ContainerStarted","Data":"2f805a0a2a40cedbc41863ae568e8e972bc38de26b1a9f1dd629dfda59b60a4c"} Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.609993 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j6gkw" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.611980 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" event={"ID":"5cb8607d-1cbd-4b3a-9b06-f2db94952d82","Type":"ContainerStarted","Data":"2ac9fd05a9335664c489379f6352d1d2002c1b227651cf5ebc79284e99e1aca3"} Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.614245 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gffs7" event={"ID":"65704ac7-fe35-4513-85b3-d512314162e8","Type":"ContainerStarted","Data":"9c3d198dc34b8ff00fda1876f89fef24206eaaa2a80c474b2bc1a814f1151e89"} Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.614328 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-gffs7" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.615924 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" event={"ID":"7bff784a-c53a-4ff1-8507-e45b2eab9f1f","Type":"ContainerStarted","Data":"80f2e5bc5bafdcc394b34c0d4e4040990a8041ae41690054a1ed88561d193467"} Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.616540 4745 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dc27t container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.616603 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" podUID="7bff784a-c53a-4ff1-8507-e45b2eab9f1f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.617373 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" event={"ID":"03110c82-b25a-4ed8-a9ae-b59bd8181baf","Type":"ContainerStarted","Data":"05477efeefbadfae37b33a9dcf20b46afca4ceb8bc8d4eca0f556bb1f990db49"} Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.621264 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-skrn4" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.621563 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-kth6t" event={"ID":"1bca2f23-df4e-4144-8253-9ad9b08df710","Type":"ContainerStarted","Data":"efd9b83e78c47866ce41d61010984a015b4316ad3e0f4a98ec886d7e02bdf54a"} Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.623619 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf" event={"ID":"a8a4af69-070d-48f1-8f56-1466adc9efb6","Type":"ContainerStarted","Data":"4e4a5a1fea3d38b0b0926c9cdcecbfd19d8b375371d4a410d0a130780ce3d712"} Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.623646 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf" event={"ID":"a8a4af69-070d-48f1-8f56-1466adc9efb6","Type":"ContainerStarted","Data":"c33152f1b336a2b811571fb176b125157573ad05e710ec3e2331a46b0d3ae5f7"} Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.625719 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-72w5j" event={"ID":"a288a2bb-c87a-464b-b2fe-d2cab9a5df70","Type":"ContainerStarted","Data":"d0c3aa0a294fe71e8485f4b7ab42fec9438c5600edc837df4a33348aeac1c2dc"} Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.628089 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xq5mm" event={"ID":"e8bcad89-0f86-4a07-9c81-aa8cac743319","Type":"ContainerStarted","Data":"2cd9251382eb8c522d4acff8f4f742b889ec9d7015baa4decb92ed2f9d4cf0c4"} Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.631149 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ncgfh" event={"ID":"c0a536ca-a2c3-4cbf-9a0d-54fb68f80510","Type":"ContainerStarted","Data":"cecbbfe13662023a1026419cf8c55edb37c8a2df994a41f91ea7910215959e6d"} Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.631690 4745 patch_prober.go:28] interesting pod/downloads-7954f5f757-7jjxn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.631767 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7jjxn" podUID="247274a0-c2f0-4150-ad4f-cf39c2003a60" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.658017 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tnbzj" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.658086 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-76fgg" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.665980 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j6gkw" podStartSLOduration=102.665958216 podStartE2EDuration="1m42.665958216s" podCreationTimestamp="2025-10-10 13:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:10.656242423 +0000 UTC m=+124.553899186" watchObservedRunningTime="2025-10-10 13:20:10.665958216 +0000 UTC m=+124.563614969" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.676402 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qs225" podStartSLOduration=103.676378029 podStartE2EDuration="1m43.676378029s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:10.112122487 +0000 UTC m=+124.009779260" watchObservedRunningTime="2025-10-10 13:20:10.676378029 +0000 UTC m=+124.574034792" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.678805 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.679127 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.679568 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:10 crc kubenswrapper[4745]: E1010 13:20:10.680002 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:11.179984991 +0000 UTC m=+125.077641754 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.680062 4745 patch_prober.go:28] interesting pod/apiserver-76f77b778f-72w5j container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.9:8443/livez\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.680088 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-72w5j" podUID="a288a2bb-c87a-464b-b2fe-d2cab9a5df70" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.9:8443/livez\": dial tcp 10.217.0.9:8443: connect: connection refused" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.706482 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-gffs7" podStartSLOduration=7.706461825 podStartE2EDuration="7.706461825s" podCreationTimestamp="2025-10-10 13:20:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:10.70450066 +0000 UTC m=+124.602157423" watchObservedRunningTime="2025-10-10 13:20:10.706461825 +0000 UTC m=+124.604118588" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.756494 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" podStartSLOduration=103.756476971 podStartE2EDuration="1m43.756476971s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:10.754857745 +0000 UTC m=+124.652514528" watchObservedRunningTime="2025-10-10 13:20:10.756476971 +0000 UTC m=+124.654133734" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.782791 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:10 crc kubenswrapper[4745]: E1010 13:20:10.799499 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:11.299468219 +0000 UTC m=+125.197125172 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.829196 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-xq5mm" podStartSLOduration=103.829168554 podStartE2EDuration="1m43.829168554s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:10.776605597 +0000 UTC m=+124.674262350" watchObservedRunningTime="2025-10-10 13:20:10.829168554 +0000 UTC m=+124.726825317" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.829974 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-ncgfh" podStartSLOduration=103.829967527 podStartE2EDuration="1m43.829967527s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:10.828323491 +0000 UTC m=+124.725980254" watchObservedRunningTime="2025-10-10 13:20:10.829967527 +0000 UTC m=+124.727624290" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.884385 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:10 crc kubenswrapper[4745]: E1010 13:20:10.884828 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:11.384804738 +0000 UTC m=+125.282461501 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.886420 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.900584 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-72w5j" podStartSLOduration=103.900555471 podStartE2EDuration="1m43.900555471s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:10.900189091 +0000 UTC m=+124.797845864" watchObservedRunningTime="2025-10-10 13:20:10.900555471 +0000 UTC m=+124.798212234" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.900885 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-kth6t" podStartSLOduration=103.90088032 podStartE2EDuration="1m43.90088032s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:10.858716785 +0000 UTC m=+124.756373548" watchObservedRunningTime="2025-10-10 13:20:10.90088032 +0000 UTC m=+124.798537083" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.929360 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mk9cf" podStartSLOduration=103.92933931 podStartE2EDuration="1m43.92933931s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:10.926103129 +0000 UTC m=+124.823759892" watchObservedRunningTime="2025-10-10 13:20:10.92933931 +0000 UTC m=+124.826996083" Oct 10 13:20:10 crc kubenswrapper[4745]: I1010 13:20:10.985446 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:10 crc kubenswrapper[4745]: E1010 13:20:10.986034 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:11.486003773 +0000 UTC m=+125.383660536 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.087288 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:11 crc kubenswrapper[4745]: E1010 13:20:11.087468 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:11.587439815 +0000 UTC m=+125.485096588 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.087931 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:11 crc kubenswrapper[4745]: E1010 13:20:11.088261 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:11.588252768 +0000 UTC m=+125.485909531 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.189136 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:11 crc kubenswrapper[4745]: E1010 13:20:11.189325 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:11.689298948 +0000 UTC m=+125.586955701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.189408 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:11 crc kubenswrapper[4745]: E1010 13:20:11.189676 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:11.689668309 +0000 UTC m=+125.587325072 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.290919 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:11 crc kubenswrapper[4745]: E1010 13:20:11.291148 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:11.79111501 +0000 UTC m=+125.688771773 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.392685 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:11 crc kubenswrapper[4745]: E1010 13:20:11.393040 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:11.893024195 +0000 UTC m=+125.790680958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.494086 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:11 crc kubenswrapper[4745]: E1010 13:20:11.494224 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:11.994207659 +0000 UTC m=+125.891864422 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.494374 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:11 crc kubenswrapper[4745]: E1010 13:20:11.494601 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:11.99459231 +0000 UTC m=+125.892249073 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.578226 4745 patch_prober.go:28] interesting pod/router-default-5444994796-5ftts container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 13:20:11 crc kubenswrapper[4745]: [-]has-synced failed: reason withheld Oct 10 13:20:11 crc kubenswrapper[4745]: [+]process-running ok Oct 10 13:20:11 crc kubenswrapper[4745]: healthz check failed Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.578280 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5ftts" podUID="f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.599221 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:11 crc kubenswrapper[4745]: E1010 13:20:11.599576 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:12.099561251 +0000 UTC m=+125.997218004 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.631649 4745 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-bmlmb container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.631718 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" podUID="4c8c7598-c1b9-4201-9f79-21c0c312f26a" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.631975 4745 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-98dn4 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.25:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.632019 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" podUID="11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.25:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.637309 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" event={"ID":"03110c82-b25a-4ed8-a9ae-b59bd8181baf","Type":"ContainerStarted","Data":"ef3222c0350b9ec5431fb37632591f9e8021eb302aa58fd161c5d73de47f0a3f"} Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.637895 4745 patch_prober.go:28] interesting pod/downloads-7954f5f757-7jjxn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.637967 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7jjxn" podUID="247274a0-c2f0-4150-ad4f-cf39c2003a60" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.638061 4745 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dc27t container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.638083 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" podUID="7bff784a-c53a-4ff1-8507-e45b2eab9f1f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.646713 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hh7hq" Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.700905 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:11 crc kubenswrapper[4745]: E1010 13:20:11.701280 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:12.20126723 +0000 UTC m=+126.098923993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.802501 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:11 crc kubenswrapper[4745]: E1010 13:20:11.804034 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:12.303987828 +0000 UTC m=+126.201644631 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.844990 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bmlmb" Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.904895 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:11 crc kubenswrapper[4745]: E1010 13:20:11.909510 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:12.409481513 +0000 UTC m=+126.307138276 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.991952 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mjlxs"] Oct 10 13:20:11 crc kubenswrapper[4745]: I1010 13:20:11.992907 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mjlxs" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.001195 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.018054 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mjlxs"] Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.021314 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:12 crc kubenswrapper[4745]: E1010 13:20:12.021598 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:12.521562154 +0000 UTC m=+126.419218917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.021713 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:12 crc kubenswrapper[4745]: E1010 13:20:12.022195 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:12.522175151 +0000 UTC m=+126.419831914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.125782 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:12 crc kubenswrapper[4745]: E1010 13:20:12.125972 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:12.625943678 +0000 UTC m=+126.523600441 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.126137 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-catalog-content\") pod \"community-operators-mjlxs\" (UID: \"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d\") " pod="openshift-marketplace/community-operators-mjlxs" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.126300 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cklnn\" (UniqueName: \"kubernetes.io/projected/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-kube-api-access-cklnn\") pod \"community-operators-mjlxs\" (UID: \"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d\") " pod="openshift-marketplace/community-operators-mjlxs" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.126392 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.126460 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-utilities\") pod \"community-operators-mjlxs\" (UID: \"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d\") " pod="openshift-marketplace/community-operators-mjlxs" Oct 10 13:20:12 crc kubenswrapper[4745]: E1010 13:20:12.126674 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:12.626665248 +0000 UTC m=+126.524322011 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.171310 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tr9k9"] Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.172128 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tr9k9" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.173676 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.227801 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:12 crc kubenswrapper[4745]: E1010 13:20:12.228074 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:12.728038308 +0000 UTC m=+126.625695071 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.228533 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.228574 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-utilities\") pod \"community-operators-mjlxs\" (UID: \"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d\") " pod="openshift-marketplace/community-operators-mjlxs" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.228629 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-catalog-content\") pod \"community-operators-mjlxs\" (UID: \"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d\") " pod="openshift-marketplace/community-operators-mjlxs" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.228699 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cklnn\" (UniqueName: \"kubernetes.io/projected/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-kube-api-access-cklnn\") pod \"community-operators-mjlxs\" (UID: \"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d\") " pod="openshift-marketplace/community-operators-mjlxs" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.229466 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-utilities\") pod \"community-operators-mjlxs\" (UID: \"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d\") " pod="openshift-marketplace/community-operators-mjlxs" Oct 10 13:20:12 crc kubenswrapper[4745]: E1010 13:20:12.229587 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:12.729572351 +0000 UTC m=+126.627229114 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.230198 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-catalog-content\") pod \"community-operators-mjlxs\" (UID: \"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d\") " pod="openshift-marketplace/community-operators-mjlxs" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.274960 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tr9k9"] Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.276325 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cklnn\" (UniqueName: \"kubernetes.io/projected/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-kube-api-access-cklnn\") pod \"community-operators-mjlxs\" (UID: \"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d\") " pod="openshift-marketplace/community-operators-mjlxs" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.311411 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mjlxs" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.329233 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:12 crc kubenswrapper[4745]: E1010 13:20:12.329411 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:12.829386887 +0000 UTC m=+126.727043650 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.329489 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44dedd6d-7915-4792-81c7-d8e2dc184478-catalog-content\") pod \"certified-operators-tr9k9\" (UID: \"44dedd6d-7915-4792-81c7-d8e2dc184478\") " pod="openshift-marketplace/certified-operators-tr9k9" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.329594 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krclq\" (UniqueName: \"kubernetes.io/projected/44dedd6d-7915-4792-81c7-d8e2dc184478-kube-api-access-krclq\") pod \"certified-operators-tr9k9\" (UID: \"44dedd6d-7915-4792-81c7-d8e2dc184478\") " pod="openshift-marketplace/certified-operators-tr9k9" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.329675 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44dedd6d-7915-4792-81c7-d8e2dc184478-utilities\") pod \"certified-operators-tr9k9\" (UID: \"44dedd6d-7915-4792-81c7-d8e2dc184478\") " pod="openshift-marketplace/certified-operators-tr9k9" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.329758 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:12 crc kubenswrapper[4745]: E1010 13:20:12.330069 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:12.830054356 +0000 UTC m=+126.727711129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.372561 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rp9wn"] Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.376757 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rp9wn" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.386029 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.405098 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rp9wn"] Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.434752 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.434992 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44dedd6d-7915-4792-81c7-d8e2dc184478-catalog-content\") pod \"certified-operators-tr9k9\" (UID: \"44dedd6d-7915-4792-81c7-d8e2dc184478\") " pod="openshift-marketplace/certified-operators-tr9k9" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.435035 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krclq\" (UniqueName: \"kubernetes.io/projected/44dedd6d-7915-4792-81c7-d8e2dc184478-kube-api-access-krclq\") pod \"certified-operators-tr9k9\" (UID: \"44dedd6d-7915-4792-81c7-d8e2dc184478\") " pod="openshift-marketplace/certified-operators-tr9k9" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.435091 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44dedd6d-7915-4792-81c7-d8e2dc184478-utilities\") pod \"certified-operators-tr9k9\" (UID: \"44dedd6d-7915-4792-81c7-d8e2dc184478\") " pod="openshift-marketplace/certified-operators-tr9k9" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.435837 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44dedd6d-7915-4792-81c7-d8e2dc184478-utilities\") pod \"certified-operators-tr9k9\" (UID: \"44dedd6d-7915-4792-81c7-d8e2dc184478\") " pod="openshift-marketplace/certified-operators-tr9k9" Oct 10 13:20:12 crc kubenswrapper[4745]: E1010 13:20:12.435893 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:12.93586514 +0000 UTC m=+126.833521903 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.436068 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44dedd6d-7915-4792-81c7-d8e2dc184478-catalog-content\") pod \"certified-operators-tr9k9\" (UID: \"44dedd6d-7915-4792-81c7-d8e2dc184478\") " pod="openshift-marketplace/certified-operators-tr9k9" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.502233 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krclq\" (UniqueName: \"kubernetes.io/projected/44dedd6d-7915-4792-81c7-d8e2dc184478-kube-api-access-krclq\") pod \"certified-operators-tr9k9\" (UID: \"44dedd6d-7915-4792-81c7-d8e2dc184478\") " pod="openshift-marketplace/certified-operators-tr9k9" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.536014 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/864139db-62ff-4c5c-8478-8cf237df2f3b-utilities\") pod \"community-operators-rp9wn\" (UID: \"864139db-62ff-4c5c-8478-8cf237df2f3b\") " pod="openshift-marketplace/community-operators-rp9wn" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.536057 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5b8x\" (UniqueName: \"kubernetes.io/projected/864139db-62ff-4c5c-8478-8cf237df2f3b-kube-api-access-h5b8x\") pod \"community-operators-rp9wn\" (UID: \"864139db-62ff-4c5c-8478-8cf237df2f3b\") " pod="openshift-marketplace/community-operators-rp9wn" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.536084 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/864139db-62ff-4c5c-8478-8cf237df2f3b-catalog-content\") pod \"community-operators-rp9wn\" (UID: \"864139db-62ff-4c5c-8478-8cf237df2f3b\") " pod="openshift-marketplace/community-operators-rp9wn" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.536113 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:12 crc kubenswrapper[4745]: E1010 13:20:12.536401 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:13.036387006 +0000 UTC m=+126.934043769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.558575 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sdzc8"] Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.559518 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sdzc8" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.561339 4745 patch_prober.go:28] interesting pod/router-default-5444994796-5ftts container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 13:20:12 crc kubenswrapper[4745]: [-]has-synced failed: reason withheld Oct 10 13:20:12 crc kubenswrapper[4745]: [+]process-running ok Oct 10 13:20:12 crc kubenswrapper[4745]: healthz check failed Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.561370 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5ftts" podUID="f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.577592 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sdzc8"] Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.636598 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.636829 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5b8x\" (UniqueName: \"kubernetes.io/projected/864139db-62ff-4c5c-8478-8cf237df2f3b-kube-api-access-h5b8x\") pod \"community-operators-rp9wn\" (UID: \"864139db-62ff-4c5c-8478-8cf237df2f3b\") " pod="openshift-marketplace/community-operators-rp9wn" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.636860 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/864139db-62ff-4c5c-8478-8cf237df2f3b-catalog-content\") pod \"community-operators-rp9wn\" (UID: \"864139db-62ff-4c5c-8478-8cf237df2f3b\") " pod="openshift-marketplace/community-operators-rp9wn" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.636969 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/864139db-62ff-4c5c-8478-8cf237df2f3b-utilities\") pod \"community-operators-rp9wn\" (UID: \"864139db-62ff-4c5c-8478-8cf237df2f3b\") " pod="openshift-marketplace/community-operators-rp9wn" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.637382 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/864139db-62ff-4c5c-8478-8cf237df2f3b-utilities\") pod \"community-operators-rp9wn\" (UID: \"864139db-62ff-4c5c-8478-8cf237df2f3b\") " pod="openshift-marketplace/community-operators-rp9wn" Oct 10 13:20:12 crc kubenswrapper[4745]: E1010 13:20:12.637446 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:13.137432366 +0000 UTC m=+127.035089129 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.637896 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/864139db-62ff-4c5c-8478-8cf237df2f3b-catalog-content\") pod \"community-operators-rp9wn\" (UID: \"864139db-62ff-4c5c-8478-8cf237df2f3b\") " pod="openshift-marketplace/community-operators-rp9wn" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.677844 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5b8x\" (UniqueName: \"kubernetes.io/projected/864139db-62ff-4c5c-8478-8cf237df2f3b-kube-api-access-h5b8x\") pod \"community-operators-rp9wn\" (UID: \"864139db-62ff-4c5c-8478-8cf237df2f3b\") " pod="openshift-marketplace/community-operators-rp9wn" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.689372 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" event={"ID":"03110c82-b25a-4ed8-a9ae-b59bd8181baf","Type":"ContainerStarted","Data":"8b0e03c2bb81b0bb6552dcfe4f6cc2eba4559630e5df2aba6cde5ad2069db930"} Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.689412 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" event={"ID":"03110c82-b25a-4ed8-a9ae-b59bd8181baf","Type":"ContainerStarted","Data":"4a05c1ba906c7ab04846ff0d1a41d5f1e4fc3abe02363f36374a22db79d6781d"} Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.711278 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rp9wn" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.715914 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-rwdj4" podStartSLOduration=9.715890812 podStartE2EDuration="9.715890812s" podCreationTimestamp="2025-10-10 13:20:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:12.715202463 +0000 UTC m=+126.612859226" watchObservedRunningTime="2025-10-10 13:20:12.715890812 +0000 UTC m=+126.613547565" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.740747 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b29ff1bb-a068-485a-891d-2270ddebd191-utilities\") pod \"certified-operators-sdzc8\" (UID: \"b29ff1bb-a068-485a-891d-2270ddebd191\") " pod="openshift-marketplace/certified-operators-sdzc8" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.740799 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.740848 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b29ff1bb-a068-485a-891d-2270ddebd191-catalog-content\") pod \"certified-operators-sdzc8\" (UID: \"b29ff1bb-a068-485a-891d-2270ddebd191\") " pod="openshift-marketplace/certified-operators-sdzc8" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.740872 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrwtn\" (UniqueName: \"kubernetes.io/projected/b29ff1bb-a068-485a-891d-2270ddebd191-kube-api-access-rrwtn\") pod \"certified-operators-sdzc8\" (UID: \"b29ff1bb-a068-485a-891d-2270ddebd191\") " pod="openshift-marketplace/certified-operators-sdzc8" Oct 10 13:20:12 crc kubenswrapper[4745]: E1010 13:20:12.741266 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:13.241253305 +0000 UTC m=+127.138910068 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.789964 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tr9k9" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.841545 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.841845 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b29ff1bb-a068-485a-891d-2270ddebd191-utilities\") pod \"certified-operators-sdzc8\" (UID: \"b29ff1bb-a068-485a-891d-2270ddebd191\") " pod="openshift-marketplace/certified-operators-sdzc8" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.841973 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b29ff1bb-a068-485a-891d-2270ddebd191-catalog-content\") pod \"certified-operators-sdzc8\" (UID: \"b29ff1bb-a068-485a-891d-2270ddebd191\") " pod="openshift-marketplace/certified-operators-sdzc8" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.842050 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrwtn\" (UniqueName: \"kubernetes.io/projected/b29ff1bb-a068-485a-891d-2270ddebd191-kube-api-access-rrwtn\") pod \"certified-operators-sdzc8\" (UID: \"b29ff1bb-a068-485a-891d-2270ddebd191\") " pod="openshift-marketplace/certified-operators-sdzc8" Oct 10 13:20:12 crc kubenswrapper[4745]: E1010 13:20:12.843784 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:13.343710895 +0000 UTC m=+127.241367658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.844130 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b29ff1bb-a068-485a-891d-2270ddebd191-catalog-content\") pod \"certified-operators-sdzc8\" (UID: \"b29ff1bb-a068-485a-891d-2270ddebd191\") " pod="openshift-marketplace/certified-operators-sdzc8" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.844356 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b29ff1bb-a068-485a-891d-2270ddebd191-utilities\") pod \"certified-operators-sdzc8\" (UID: \"b29ff1bb-a068-485a-891d-2270ddebd191\") " pod="openshift-marketplace/certified-operators-sdzc8" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.874463 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrwtn\" (UniqueName: \"kubernetes.io/projected/b29ff1bb-a068-485a-891d-2270ddebd191-kube-api-access-rrwtn\") pod \"certified-operators-sdzc8\" (UID: \"b29ff1bb-a068-485a-891d-2270ddebd191\") " pod="openshift-marketplace/certified-operators-sdzc8" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.929638 4745 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.934493 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sdzc8" Oct 10 13:20:12 crc kubenswrapper[4745]: I1010 13:20:12.944369 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:12 crc kubenswrapper[4745]: E1010 13:20:12.944936 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:13.44492122 +0000 UTC m=+127.342577983 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.046038 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:13 crc kubenswrapper[4745]: E1010 13:20:13.046162 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:13.546137736 +0000 UTC m=+127.443794499 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.046431 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:13 crc kubenswrapper[4745]: E1010 13:20:13.046843 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:13.546834735 +0000 UTC m=+127.444491498 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.047395 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mjlxs"] Oct 10 13:20:13 crc kubenswrapper[4745]: W1010 13:20:13.088994 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0dbfa2cf_9426_41e9_9f46_1cfc39e2210d.slice/crio-c8fdf7017342861a13d882ac126123dce706c30e5872f89277bcf693a92951be WatchSource:0}: Error finding container c8fdf7017342861a13d882ac126123dce706c30e5872f89277bcf693a92951be: Status 404 returned error can't find the container with id c8fdf7017342861a13d882ac126123dce706c30e5872f89277bcf693a92951be Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.150341 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:13 crc kubenswrapper[4745]: E1010 13:20:13.151158 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:13.651132257 +0000 UTC m=+127.548789110 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.151236 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:13 crc kubenswrapper[4745]: E1010 13:20:13.151556 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:13.651549629 +0000 UTC m=+127.549206392 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.201520 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rp9wn"] Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.251915 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:13 crc kubenswrapper[4745]: E1010 13:20:13.252262 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:13.75224544 +0000 UTC m=+127.649902203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.310685 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tr9k9"] Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.354370 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:13 crc kubenswrapper[4745]: E1010 13:20:13.355004 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:13.854976928 +0000 UTC m=+127.752633691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.382996 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sdzc8"] Oct 10 13:20:13 crc kubenswrapper[4745]: W1010 13:20:13.405212 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb29ff1bb_a068_485a_891d_2270ddebd191.slice/crio-c76ef6884d858922712c8592c0fcf6729f0c738bd260253f63022bb5572594f6 WatchSource:0}: Error finding container c76ef6884d858922712c8592c0fcf6729f0c738bd260253f63022bb5572594f6: Status 404 returned error can't find the container with id c76ef6884d858922712c8592c0fcf6729f0c738bd260253f63022bb5572594f6 Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.455158 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:13 crc kubenswrapper[4745]: E1010 13:20:13.455416 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:13.955383999 +0000 UTC m=+127.853040762 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.455499 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:13 crc kubenswrapper[4745]: E1010 13:20:13.455970 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:13.955963596 +0000 UTC m=+127.853620349 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.556214 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:13 crc kubenswrapper[4745]: E1010 13:20:13.556524 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:14.056472691 +0000 UTC m=+127.954129464 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.556687 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:13 crc kubenswrapper[4745]: E1010 13:20:13.557048 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:14.057030557 +0000 UTC m=+127.954687320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.573547 4745 patch_prober.go:28] interesting pod/router-default-5444994796-5ftts container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 13:20:13 crc kubenswrapper[4745]: [-]has-synced failed: reason withheld Oct 10 13:20:13 crc kubenswrapper[4745]: [+]process-running ok Oct 10 13:20:13 crc kubenswrapper[4745]: healthz check failed Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.573631 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5ftts" podUID="f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.657445 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:13 crc kubenswrapper[4745]: E1010 13:20:13.657626 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:14.157596194 +0000 UTC m=+128.055252957 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.657810 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:13 crc kubenswrapper[4745]: E1010 13:20:13.658161 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:14.158145079 +0000 UTC m=+128.055801952 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.694138 4745 generic.go:334] "Generic (PLEG): container finished" podID="0dbfa2cf-9426-41e9-9f46-1cfc39e2210d" containerID="7c12c75367c2ec55e7a4fb89940a11df819f4e956a445de25d20773bc586438b" exitCode=0 Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.694217 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjlxs" event={"ID":"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d","Type":"ContainerDied","Data":"7c12c75367c2ec55e7a4fb89940a11df819f4e956a445de25d20773bc586438b"} Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.694266 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjlxs" event={"ID":"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d","Type":"ContainerStarted","Data":"c8fdf7017342861a13d882ac126123dce706c30e5872f89277bcf693a92951be"} Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.696335 4745 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.696683 4745 generic.go:334] "Generic (PLEG): container finished" podID="44dedd6d-7915-4792-81c7-d8e2dc184478" containerID="3d37cb7176fc1f6e218e9c36d9663913ea1f40a50e464386ff304905f119272d" exitCode=0 Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.696744 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tr9k9" event={"ID":"44dedd6d-7915-4792-81c7-d8e2dc184478","Type":"ContainerDied","Data":"3d37cb7176fc1f6e218e9c36d9663913ea1f40a50e464386ff304905f119272d"} Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.696767 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tr9k9" event={"ID":"44dedd6d-7915-4792-81c7-d8e2dc184478","Type":"ContainerStarted","Data":"e3f76a1b7fc7670e96525f9edc699918cb36bf5e9864e3241f3f15b77ff17494"} Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.705181 4745 generic.go:334] "Generic (PLEG): container finished" podID="b29ff1bb-a068-485a-891d-2270ddebd191" containerID="46f975de84b406276106bc9e1135fea1ff94946bd0882d033fc240490b1a7dd8" exitCode=0 Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.705289 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sdzc8" event={"ID":"b29ff1bb-a068-485a-891d-2270ddebd191","Type":"ContainerDied","Data":"46f975de84b406276106bc9e1135fea1ff94946bd0882d033fc240490b1a7dd8"} Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.705340 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sdzc8" event={"ID":"b29ff1bb-a068-485a-891d-2270ddebd191","Type":"ContainerStarted","Data":"c76ef6884d858922712c8592c0fcf6729f0c738bd260253f63022bb5572594f6"} Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.707830 4745 generic.go:334] "Generic (PLEG): container finished" podID="864139db-62ff-4c5c-8478-8cf237df2f3b" containerID="7a5f4460de68703a5b6619368deeb9bb92506dcebbf0d9c0d3d92488bdc76531" exitCode=0 Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.707950 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rp9wn" event={"ID":"864139db-62ff-4c5c-8478-8cf237df2f3b","Type":"ContainerDied","Data":"7a5f4460de68703a5b6619368deeb9bb92506dcebbf0d9c0d3d92488bdc76531"} Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.707990 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rp9wn" event={"ID":"864139db-62ff-4c5c-8478-8cf237df2f3b","Type":"ContainerStarted","Data":"ead87654d18f0dd9bc336113704ca06da965ec8fff7206a5fe94c8ad805bbdcd"} Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.759095 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:13 crc kubenswrapper[4745]: E1010 13:20:13.759337 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:14.259294762 +0000 UTC m=+128.156951525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.759684 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:13 crc kubenswrapper[4745]: E1010 13:20:13.759995 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:14.259980932 +0000 UTC m=+128.157637685 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.784329 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.785051 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.790567 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.793801 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.794430 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.860494 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:13 crc kubenswrapper[4745]: E1010 13:20:13.860893 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 13:20:14.360863378 +0000 UTC m=+128.258520151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.861121 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9547e29a-9de8-42c8-bc91-8366a138d1fc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9547e29a-9de8-42c8-bc91-8366a138d1fc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.861286 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9547e29a-9de8-42c8-bc91-8366a138d1fc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9547e29a-9de8-42c8-bc91-8366a138d1fc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.861404 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:13 crc kubenswrapper[4745]: E1010 13:20:13.861763 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 13:20:14.361752023 +0000 UTC m=+128.259408796 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qss59" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.906544 4745 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-10T13:20:12.929678732Z","Handler":null,"Name":""} Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.915859 4745 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.915911 4745 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.962542 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.963060 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9547e29a-9de8-42c8-bc91-8366a138d1fc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9547e29a-9de8-42c8-bc91-8366a138d1fc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.963123 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9547e29a-9de8-42c8-bc91-8366a138d1fc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9547e29a-9de8-42c8-bc91-8366a138d1fc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.963254 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9547e29a-9de8-42c8-bc91-8366a138d1fc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9547e29a-9de8-42c8-bc91-8366a138d1fc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.968581 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 10 13:20:13 crc kubenswrapper[4745]: I1010 13:20:13.983427 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9547e29a-9de8-42c8-bc91-8366a138d1fc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9547e29a-9de8-42c8-bc91-8366a138d1fc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.064451 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.067462 4745 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.067585 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.100984 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qss59\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.101380 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.138049 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.147492 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.163841 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tb4mp"] Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.164853 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tb4mp" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.170937 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.173533 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tb4mp"] Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.270564 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8674419-b98d-4f7e-ae45-b20aca88dcd5-utilities\") pod \"redhat-marketplace-tb4mp\" (UID: \"c8674419-b98d-4f7e-ae45-b20aca88dcd5\") " pod="openshift-marketplace/redhat-marketplace-tb4mp" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.270625 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddcwd\" (UniqueName: \"kubernetes.io/projected/c8674419-b98d-4f7e-ae45-b20aca88dcd5-kube-api-access-ddcwd\") pod \"redhat-marketplace-tb4mp\" (UID: \"c8674419-b98d-4f7e-ae45-b20aca88dcd5\") " pod="openshift-marketplace/redhat-marketplace-tb4mp" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.270648 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8674419-b98d-4f7e-ae45-b20aca88dcd5-catalog-content\") pod \"redhat-marketplace-tb4mp\" (UID: \"c8674419-b98d-4f7e-ae45-b20aca88dcd5\") " pod="openshift-marketplace/redhat-marketplace-tb4mp" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.358171 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.372778 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8674419-b98d-4f7e-ae45-b20aca88dcd5-catalog-content\") pod \"redhat-marketplace-tb4mp\" (UID: \"c8674419-b98d-4f7e-ae45-b20aca88dcd5\") " pod="openshift-marketplace/redhat-marketplace-tb4mp" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.372977 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8674419-b98d-4f7e-ae45-b20aca88dcd5-utilities\") pod \"redhat-marketplace-tb4mp\" (UID: \"c8674419-b98d-4f7e-ae45-b20aca88dcd5\") " pod="openshift-marketplace/redhat-marketplace-tb4mp" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.373067 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddcwd\" (UniqueName: \"kubernetes.io/projected/c8674419-b98d-4f7e-ae45-b20aca88dcd5-kube-api-access-ddcwd\") pod \"redhat-marketplace-tb4mp\" (UID: \"c8674419-b98d-4f7e-ae45-b20aca88dcd5\") " pod="openshift-marketplace/redhat-marketplace-tb4mp" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.373233 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8674419-b98d-4f7e-ae45-b20aca88dcd5-catalog-content\") pod \"redhat-marketplace-tb4mp\" (UID: \"c8674419-b98d-4f7e-ae45-b20aca88dcd5\") " pod="openshift-marketplace/redhat-marketplace-tb4mp" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.373843 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8674419-b98d-4f7e-ae45-b20aca88dcd5-utilities\") pod \"redhat-marketplace-tb4mp\" (UID: \"c8674419-b98d-4f7e-ae45-b20aca88dcd5\") " pod="openshift-marketplace/redhat-marketplace-tb4mp" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.391617 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddcwd\" (UniqueName: \"kubernetes.io/projected/c8674419-b98d-4f7e-ae45-b20aca88dcd5-kube-api-access-ddcwd\") pod \"redhat-marketplace-tb4mp\" (UID: \"c8674419-b98d-4f7e-ae45-b20aca88dcd5\") " pod="openshift-marketplace/redhat-marketplace-tb4mp" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.426778 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qss59"] Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.500712 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tb4mp" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.571247 4745 patch_prober.go:28] interesting pod/router-default-5444994796-5ftts container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 13:20:14 crc kubenswrapper[4745]: [-]has-synced failed: reason withheld Oct 10 13:20:14 crc kubenswrapper[4745]: [+]process-running ok Oct 10 13:20:14 crc kubenswrapper[4745]: healthz check failed Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.571327 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5ftts" podUID="f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.577690 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bbpdf"] Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.579893 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bbpdf" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.583765 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bbpdf"] Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.676352 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15adca8d-0d7e-4793-a3f6-489d3aa12106-catalog-content\") pod \"redhat-marketplace-bbpdf\" (UID: \"15adca8d-0d7e-4793-a3f6-489d3aa12106\") " pod="openshift-marketplace/redhat-marketplace-bbpdf" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.676923 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15adca8d-0d7e-4793-a3f6-489d3aa12106-utilities\") pod \"redhat-marketplace-bbpdf\" (UID: \"15adca8d-0d7e-4793-a3f6-489d3aa12106\") " pod="openshift-marketplace/redhat-marketplace-bbpdf" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.677079 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghtwr\" (UniqueName: \"kubernetes.io/projected/15adca8d-0d7e-4793-a3f6-489d3aa12106-kube-api-access-ghtwr\") pod \"redhat-marketplace-bbpdf\" (UID: \"15adca8d-0d7e-4793-a3f6-489d3aa12106\") " pod="openshift-marketplace/redhat-marketplace-bbpdf" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.718292 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9547e29a-9de8-42c8-bc91-8366a138d1fc","Type":"ContainerStarted","Data":"35fb716b09be07bf5c21ce1b8d951ca5e80ecc929bc2c1ffa435d7bc023139b4"} Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.720462 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qss59" event={"ID":"354f8f3e-647a-4be3-a2c1-26b2cdb336b1","Type":"ContainerStarted","Data":"e667f4bfe244728ad6f95d1940e67643e4168315b728acb023d895a2a8ddf503"} Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.720502 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qss59" event={"ID":"354f8f3e-647a-4be3-a2c1-26b2cdb336b1","Type":"ContainerStarted","Data":"cd7b478275debc46ce04591111b8c4dde0276fb0f496777e745834d5df7ce6c8"} Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.721923 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.748160 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-qss59" podStartSLOduration=107.74814331 podStartE2EDuration="1m47.74814331s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:14.744986091 +0000 UTC m=+128.642642854" watchObservedRunningTime="2025-10-10 13:20:14.74814331 +0000 UTC m=+128.645800073" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.763857 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.779033 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghtwr\" (UniqueName: \"kubernetes.io/projected/15adca8d-0d7e-4793-a3f6-489d3aa12106-kube-api-access-ghtwr\") pod \"redhat-marketplace-bbpdf\" (UID: \"15adca8d-0d7e-4793-a3f6-489d3aa12106\") " pod="openshift-marketplace/redhat-marketplace-bbpdf" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.779166 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15adca8d-0d7e-4793-a3f6-489d3aa12106-catalog-content\") pod \"redhat-marketplace-bbpdf\" (UID: \"15adca8d-0d7e-4793-a3f6-489d3aa12106\") " pod="openshift-marketplace/redhat-marketplace-bbpdf" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.779200 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15adca8d-0d7e-4793-a3f6-489d3aa12106-utilities\") pod \"redhat-marketplace-bbpdf\" (UID: \"15adca8d-0d7e-4793-a3f6-489d3aa12106\") " pod="openshift-marketplace/redhat-marketplace-bbpdf" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.779610 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15adca8d-0d7e-4793-a3f6-489d3aa12106-utilities\") pod \"redhat-marketplace-bbpdf\" (UID: \"15adca8d-0d7e-4793-a3f6-489d3aa12106\") " pod="openshift-marketplace/redhat-marketplace-bbpdf" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.780236 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15adca8d-0d7e-4793-a3f6-489d3aa12106-catalog-content\") pod \"redhat-marketplace-bbpdf\" (UID: \"15adca8d-0d7e-4793-a3f6-489d3aa12106\") " pod="openshift-marketplace/redhat-marketplace-bbpdf" Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.796286 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tb4mp"] Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.801310 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghtwr\" (UniqueName: \"kubernetes.io/projected/15adca8d-0d7e-4793-a3f6-489d3aa12106-kube-api-access-ghtwr\") pod \"redhat-marketplace-bbpdf\" (UID: \"15adca8d-0d7e-4793-a3f6-489d3aa12106\") " pod="openshift-marketplace/redhat-marketplace-bbpdf" Oct 10 13:20:14 crc kubenswrapper[4745]: W1010 13:20:14.805822 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8674419_b98d_4f7e_ae45_b20aca88dcd5.slice/crio-42109b9143c8be02196abe79d359bfafe4b0d9290ff470fac9333db657d89305 WatchSource:0}: Error finding container 42109b9143c8be02196abe79d359bfafe4b0d9290ff470fac9333db657d89305: Status 404 returned error can't find the container with id 42109b9143c8be02196abe79d359bfafe4b0d9290ff470fac9333db657d89305 Oct 10 13:20:14 crc kubenswrapper[4745]: I1010 13:20:14.914500 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bbpdf" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.156624 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7vxfb"] Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.158407 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7vxfb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.164985 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.173813 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7vxfb"] Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.183585 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-catalog-content\") pod \"redhat-operators-7vxfb\" (UID: \"0dd61fe6-4685-4ce4-9c35-6465d1aa666c\") " pod="openshift-marketplace/redhat-operators-7vxfb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.183630 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff27t\" (UniqueName: \"kubernetes.io/projected/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-kube-api-access-ff27t\") pod \"redhat-operators-7vxfb\" (UID: \"0dd61fe6-4685-4ce4-9c35-6465d1aa666c\") " pod="openshift-marketplace/redhat-operators-7vxfb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.183809 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-utilities\") pod \"redhat-operators-7vxfb\" (UID: \"0dd61fe6-4685-4ce4-9c35-6465d1aa666c\") " pod="openshift-marketplace/redhat-operators-7vxfb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.187853 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bbpdf"] Oct 10 13:20:15 crc kubenswrapper[4745]: W1010 13:20:15.214148 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15adca8d_0d7e_4793_a3f6_489d3aa12106.slice/crio-e42ce376fcacc2b2ee434f439139cfb9cb33cd002e0c53212777d701a251194b WatchSource:0}: Error finding container e42ce376fcacc2b2ee434f439139cfb9cb33cd002e0c53212777d701a251194b: Status 404 returned error can't find the container with id e42ce376fcacc2b2ee434f439139cfb9cb33cd002e0c53212777d701a251194b Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.285021 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-catalog-content\") pod \"redhat-operators-7vxfb\" (UID: \"0dd61fe6-4685-4ce4-9c35-6465d1aa666c\") " pod="openshift-marketplace/redhat-operators-7vxfb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.285099 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff27t\" (UniqueName: \"kubernetes.io/projected/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-kube-api-access-ff27t\") pod \"redhat-operators-7vxfb\" (UID: \"0dd61fe6-4685-4ce4-9c35-6465d1aa666c\") " pod="openshift-marketplace/redhat-operators-7vxfb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.285169 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-utilities\") pod \"redhat-operators-7vxfb\" (UID: \"0dd61fe6-4685-4ce4-9c35-6465d1aa666c\") " pod="openshift-marketplace/redhat-operators-7vxfb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.285692 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-catalog-content\") pod \"redhat-operators-7vxfb\" (UID: \"0dd61fe6-4685-4ce4-9c35-6465d1aa666c\") " pod="openshift-marketplace/redhat-operators-7vxfb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.285716 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-utilities\") pod \"redhat-operators-7vxfb\" (UID: \"0dd61fe6-4685-4ce4-9c35-6465d1aa666c\") " pod="openshift-marketplace/redhat-operators-7vxfb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.301480 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff27t\" (UniqueName: \"kubernetes.io/projected/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-kube-api-access-ff27t\") pod \"redhat-operators-7vxfb\" (UID: \"0dd61fe6-4685-4ce4-9c35-6465d1aa666c\") " pod="openshift-marketplace/redhat-operators-7vxfb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.487961 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7vxfb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.562209 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.568503 4745 patch_prober.go:28] interesting pod/router-default-5444994796-5ftts container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 13:20:15 crc kubenswrapper[4745]: [-]has-synced failed: reason withheld Oct 10 13:20:15 crc kubenswrapper[4745]: [+]process-running ok Oct 10 13:20:15 crc kubenswrapper[4745]: healthz check failed Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.568539 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5ftts" podUID="f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.572621 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2k6hb"] Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.573834 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2k6hb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.575265 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2k6hb"] Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.589612 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fa4904c-594a-4a01-aa75-e20c1ebd88af-catalog-content\") pod \"redhat-operators-2k6hb\" (UID: \"2fa4904c-594a-4a01-aa75-e20c1ebd88af\") " pod="openshift-marketplace/redhat-operators-2k6hb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.589660 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsprc\" (UniqueName: \"kubernetes.io/projected/2fa4904c-594a-4a01-aa75-e20c1ebd88af-kube-api-access-jsprc\") pod \"redhat-operators-2k6hb\" (UID: \"2fa4904c-594a-4a01-aa75-e20c1ebd88af\") " pod="openshift-marketplace/redhat-operators-2k6hb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.589699 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fa4904c-594a-4a01-aa75-e20c1ebd88af-utilities\") pod \"redhat-operators-2k6hb\" (UID: \"2fa4904c-594a-4a01-aa75-e20c1ebd88af\") " pod="openshift-marketplace/redhat-operators-2k6hb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.682793 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.692384 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsprc\" (UniqueName: \"kubernetes.io/projected/2fa4904c-594a-4a01-aa75-e20c1ebd88af-kube-api-access-jsprc\") pod \"redhat-operators-2k6hb\" (UID: \"2fa4904c-594a-4a01-aa75-e20c1ebd88af\") " pod="openshift-marketplace/redhat-operators-2k6hb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.692460 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fa4904c-594a-4a01-aa75-e20c1ebd88af-utilities\") pod \"redhat-operators-2k6hb\" (UID: \"2fa4904c-594a-4a01-aa75-e20c1ebd88af\") " pod="openshift-marketplace/redhat-operators-2k6hb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.692603 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fa4904c-594a-4a01-aa75-e20c1ebd88af-catalog-content\") pod \"redhat-operators-2k6hb\" (UID: \"2fa4904c-594a-4a01-aa75-e20c1ebd88af\") " pod="openshift-marketplace/redhat-operators-2k6hb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.692997 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fa4904c-594a-4a01-aa75-e20c1ebd88af-catalog-content\") pod \"redhat-operators-2k6hb\" (UID: \"2fa4904c-594a-4a01-aa75-e20c1ebd88af\") " pod="openshift-marketplace/redhat-operators-2k6hb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.695905 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-72w5j" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.696504 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fa4904c-594a-4a01-aa75-e20c1ebd88af-utilities\") pod \"redhat-operators-2k6hb\" (UID: \"2fa4904c-594a-4a01-aa75-e20c1ebd88af\") " pod="openshift-marketplace/redhat-operators-2k6hb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.727822 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsprc\" (UniqueName: \"kubernetes.io/projected/2fa4904c-594a-4a01-aa75-e20c1ebd88af-kube-api-access-jsprc\") pod \"redhat-operators-2k6hb\" (UID: \"2fa4904c-594a-4a01-aa75-e20c1ebd88af\") " pod="openshift-marketplace/redhat-operators-2k6hb" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.766306 4745 patch_prober.go:28] interesting pod/downloads-7954f5f757-7jjxn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.766356 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7jjxn" podUID="247274a0-c2f0-4150-ad4f-cf39c2003a60" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.766626 4745 patch_prober.go:28] interesting pod/downloads-7954f5f757-7jjxn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.766641 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-7jjxn" podUID="247274a0-c2f0-4150-ad4f-cf39c2003a60" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.778693 4745 generic.go:334] "Generic (PLEG): container finished" podID="9547e29a-9de8-42c8-bc91-8366a138d1fc" containerID="50d800ee53ddd7ddefaa3829de8e98c3187cdfe959d18527af80013ffcdcf4c9" exitCode=0 Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.778773 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9547e29a-9de8-42c8-bc91-8366a138d1fc","Type":"ContainerDied","Data":"50d800ee53ddd7ddefaa3829de8e98c3187cdfe959d18527af80013ffcdcf4c9"} Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.801022 4745 generic.go:334] "Generic (PLEG): container finished" podID="c8674419-b98d-4f7e-ae45-b20aca88dcd5" containerID="e41a4172924e16e15fcb393fb53b117ef007cb4fa979524b57b0cc12af705814" exitCode=0 Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.801122 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tb4mp" event={"ID":"c8674419-b98d-4f7e-ae45-b20aca88dcd5","Type":"ContainerDied","Data":"e41a4172924e16e15fcb393fb53b117ef007cb4fa979524b57b0cc12af705814"} Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.801155 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tb4mp" event={"ID":"c8674419-b98d-4f7e-ae45-b20aca88dcd5","Type":"ContainerStarted","Data":"42109b9143c8be02196abe79d359bfafe4b0d9290ff470fac9333db657d89305"} Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.805222 4745 generic.go:334] "Generic (PLEG): container finished" podID="15adca8d-0d7e-4793-a3f6-489d3aa12106" containerID="191efae0e5d2efb79650d7a64835385de626bd322644f6646e08d3c890ee94f0" exitCode=0 Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.806296 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbpdf" event={"ID":"15adca8d-0d7e-4793-a3f6-489d3aa12106","Type":"ContainerDied","Data":"191efae0e5d2efb79650d7a64835385de626bd322644f6646e08d3c890ee94f0"} Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.806319 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbpdf" event={"ID":"15adca8d-0d7e-4793-a3f6-489d3aa12106","Type":"ContainerStarted","Data":"e42ce376fcacc2b2ee434f439139cfb9cb33cd002e0c53212777d701a251194b"} Oct 10 13:20:15 crc kubenswrapper[4745]: I1010 13:20:15.893057 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2k6hb" Oct 10 13:20:16 crc kubenswrapper[4745]: I1010 13:20:16.072223 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7vxfb"] Oct 10 13:20:16 crc kubenswrapper[4745]: I1010 13:20:16.240259 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2k6hb"] Oct 10 13:20:16 crc kubenswrapper[4745]: I1010 13:20:16.257714 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" Oct 10 13:20:16 crc kubenswrapper[4745]: I1010 13:20:16.504696 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:16 crc kubenswrapper[4745]: I1010 13:20:16.505149 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:16 crc kubenswrapper[4745]: I1010 13:20:16.506236 4745 patch_prober.go:28] interesting pod/console-f9d7485db-xq5mm container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.35:8443/health\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Oct 10 13:20:16 crc kubenswrapper[4745]: I1010 13:20:16.506290 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-xq5mm" podUID="e8bcad89-0f86-4a07-9c81-aa8cac743319" containerName="console" probeResult="failure" output="Get \"https://10.217.0.35:8443/health\": dial tcp 10.217.0.35:8443: connect: connection refused" Oct 10 13:20:16 crc kubenswrapper[4745]: I1010 13:20:16.562240 4745 patch_prober.go:28] interesting pod/router-default-5444994796-5ftts container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 13:20:16 crc kubenswrapper[4745]: [-]has-synced failed: reason withheld Oct 10 13:20:16 crc kubenswrapper[4745]: [+]process-running ok Oct 10 13:20:16 crc kubenswrapper[4745]: healthz check failed Oct 10 13:20:16 crc kubenswrapper[4745]: I1010 13:20:16.562289 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5ftts" podUID="f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 13:20:16 crc kubenswrapper[4745]: I1010 13:20:16.816817 4745 generic.go:334] "Generic (PLEG): container finished" podID="5cb8607d-1cbd-4b3a-9b06-f2db94952d82" containerID="2ac9fd05a9335664c489379f6352d1d2002c1b227651cf5ebc79284e99e1aca3" exitCode=0 Oct 10 13:20:16 crc kubenswrapper[4745]: I1010 13:20:16.816921 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" event={"ID":"5cb8607d-1cbd-4b3a-9b06-f2db94952d82","Type":"ContainerDied","Data":"2ac9fd05a9335664c489379f6352d1d2002c1b227651cf5ebc79284e99e1aca3"} Oct 10 13:20:16 crc kubenswrapper[4745]: I1010 13:20:16.821227 4745 generic.go:334] "Generic (PLEG): container finished" podID="0dd61fe6-4685-4ce4-9c35-6465d1aa666c" containerID="0d61d2539a10e78356994e3119e9f5223131f138e460508848bebb59ca5314e8" exitCode=0 Oct 10 13:20:16 crc kubenswrapper[4745]: I1010 13:20:16.821293 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vxfb" event={"ID":"0dd61fe6-4685-4ce4-9c35-6465d1aa666c","Type":"ContainerDied","Data":"0d61d2539a10e78356994e3119e9f5223131f138e460508848bebb59ca5314e8"} Oct 10 13:20:16 crc kubenswrapper[4745]: I1010 13:20:16.821319 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vxfb" event={"ID":"0dd61fe6-4685-4ce4-9c35-6465d1aa666c","Type":"ContainerStarted","Data":"2cb0c72220cae02f3c289ec548884e64429e7921bbb7ecc07ef4d11da7c20b3b"} Oct 10 13:20:16 crc kubenswrapper[4745]: I1010 13:20:16.827869 4745 generic.go:334] "Generic (PLEG): container finished" podID="2fa4904c-594a-4a01-aa75-e20c1ebd88af" containerID="decb9e8b0ad4f92623615f0efd7ea7790efac205503de47088372c1ce3318556" exitCode=0 Oct 10 13:20:16 crc kubenswrapper[4745]: I1010 13:20:16.827955 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2k6hb" event={"ID":"2fa4904c-594a-4a01-aa75-e20c1ebd88af","Type":"ContainerDied","Data":"decb9e8b0ad4f92623615f0efd7ea7790efac205503de47088372c1ce3318556"} Oct 10 13:20:16 crc kubenswrapper[4745]: I1010 13:20:16.827992 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2k6hb" event={"ID":"2fa4904c-594a-4a01-aa75-e20c1ebd88af","Type":"ContainerStarted","Data":"bbb094c761ad0989c623cd7a86ed17431b16dcdde290dfbd8eaed91aea116523"} Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.116919 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.218938 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9547e29a-9de8-42c8-bc91-8366a138d1fc-kube-api-access\") pod \"9547e29a-9de8-42c8-bc91-8366a138d1fc\" (UID: \"9547e29a-9de8-42c8-bc91-8366a138d1fc\") " Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.219395 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9547e29a-9de8-42c8-bc91-8366a138d1fc-kubelet-dir\") pod \"9547e29a-9de8-42c8-bc91-8366a138d1fc\" (UID: \"9547e29a-9de8-42c8-bc91-8366a138d1fc\") " Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.219852 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9547e29a-9de8-42c8-bc91-8366a138d1fc-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9547e29a-9de8-42c8-bc91-8366a138d1fc" (UID: "9547e29a-9de8-42c8-bc91-8366a138d1fc"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.241264 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9547e29a-9de8-42c8-bc91-8366a138d1fc-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9547e29a-9de8-42c8-bc91-8366a138d1fc" (UID: "9547e29a-9de8-42c8-bc91-8366a138d1fc"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.321185 4745 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9547e29a-9de8-42c8-bc91-8366a138d1fc-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.321213 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9547e29a-9de8-42c8-bc91-8366a138d1fc-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.569084 4745 patch_prober.go:28] interesting pod/router-default-5444994796-5ftts container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 13:20:17 crc kubenswrapper[4745]: [-]has-synced failed: reason withheld Oct 10 13:20:17 crc kubenswrapper[4745]: [+]process-running ok Oct 10 13:20:17 crc kubenswrapper[4745]: healthz check failed Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.569146 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5ftts" podUID="f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.849436 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.857637 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9547e29a-9de8-42c8-bc91-8366a138d1fc","Type":"ContainerDied","Data":"35fb716b09be07bf5c21ce1b8d951ca5e80ecc929bc2c1ffa435d7bc023139b4"} Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.857688 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35fb716b09be07bf5c21ce1b8d951ca5e80ecc929bc2c1ffa435d7bc023139b4" Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.889215 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 10 13:20:17 crc kubenswrapper[4745]: E1010 13:20:17.891788 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9547e29a-9de8-42c8-bc91-8366a138d1fc" containerName="pruner" Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.891812 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="9547e29a-9de8-42c8-bc91-8366a138d1fc" containerName="pruner" Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.891918 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="9547e29a-9de8-42c8-bc91-8366a138d1fc" containerName="pruner" Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.892313 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.894922 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.894966 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 10 13:20:17 crc kubenswrapper[4745]: I1010 13:20:17.914228 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.039828 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-gffs7" Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.064365 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.064467 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.165791 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.165873 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.166818 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.206097 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.254515 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.565267 4745 patch_prober.go:28] interesting pod/router-default-5444994796-5ftts container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 13:20:18 crc kubenswrapper[4745]: [-]has-synced failed: reason withheld Oct 10 13:20:18 crc kubenswrapper[4745]: [+]process-running ok Oct 10 13:20:18 crc kubenswrapper[4745]: healthz check failed Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.565317 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5ftts" podUID="f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.570297 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.673648 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-config-volume\") pod \"5cb8607d-1cbd-4b3a-9b06-f2db94952d82\" (UID: \"5cb8607d-1cbd-4b3a-9b06-f2db94952d82\") " Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.674114 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9nxm\" (UniqueName: \"kubernetes.io/projected/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-kube-api-access-l9nxm\") pod \"5cb8607d-1cbd-4b3a-9b06-f2db94952d82\" (UID: \"5cb8607d-1cbd-4b3a-9b06-f2db94952d82\") " Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.674328 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-secret-volume\") pod \"5cb8607d-1cbd-4b3a-9b06-f2db94952d82\" (UID: \"5cb8607d-1cbd-4b3a-9b06-f2db94952d82\") " Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.675290 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-config-volume" (OuterVolumeSpecName: "config-volume") pod "5cb8607d-1cbd-4b3a-9b06-f2db94952d82" (UID: "5cb8607d-1cbd-4b3a-9b06-f2db94952d82"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.675524 4745 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.679625 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-kube-api-access-l9nxm" (OuterVolumeSpecName: "kube-api-access-l9nxm") pod "5cb8607d-1cbd-4b3a-9b06-f2db94952d82" (UID: "5cb8607d-1cbd-4b3a-9b06-f2db94952d82"). InnerVolumeSpecName "kube-api-access-l9nxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.679935 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5cb8607d-1cbd-4b3a-9b06-f2db94952d82" (UID: "5cb8607d-1cbd-4b3a-9b06-f2db94952d82"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.777184 4745 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.777221 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9nxm\" (UniqueName: \"kubernetes.io/projected/5cb8607d-1cbd-4b3a-9b06-f2db94952d82-kube-api-access-l9nxm\") on node \"crc\" DevicePath \"\"" Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.829855 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 10 13:20:18 crc kubenswrapper[4745]: W1010 13:20:18.850115 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod8051b8a9_7a9c_4fb8_83d7_8067fd8ab4d5.slice/crio-a7b5abd44f459c4bcfdb56b93317a67002ed6f99348a56564bf85d776944f497 WatchSource:0}: Error finding container a7b5abd44f459c4bcfdb56b93317a67002ed6f99348a56564bf85d776944f497: Status 404 returned error can't find the container with id a7b5abd44f459c4bcfdb56b93317a67002ed6f99348a56564bf85d776944f497 Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.870350 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" event={"ID":"5cb8607d-1cbd-4b3a-9b06-f2db94952d82","Type":"ContainerDied","Data":"b17eb1972d0373c465ad9fdf8d4491f220a3eda08881637d0874b1713d219818"} Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.870388 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b17eb1972d0373c465ad9fdf8d4491f220a3eda08881637d0874b1713d219818" Oct 10 13:20:18 crc kubenswrapper[4745]: I1010 13:20:18.870440 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w" Oct 10 13:20:19 crc kubenswrapper[4745]: I1010 13:20:19.563857 4745 patch_prober.go:28] interesting pod/router-default-5444994796-5ftts container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 13:20:19 crc kubenswrapper[4745]: [-]has-synced failed: reason withheld Oct 10 13:20:19 crc kubenswrapper[4745]: [+]process-running ok Oct 10 13:20:19 crc kubenswrapper[4745]: healthz check failed Oct 10 13:20:19 crc kubenswrapper[4745]: I1010 13:20:19.564149 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5ftts" podUID="f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 13:20:19 crc kubenswrapper[4745]: I1010 13:20:19.963667 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5","Type":"ContainerStarted","Data":"a7b5abd44f459c4bcfdb56b93317a67002ed6f99348a56564bf85d776944f497"} Oct 10 13:20:20 crc kubenswrapper[4745]: I1010 13:20:20.411443 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:20:20 crc kubenswrapper[4745]: I1010 13:20:20.561119 4745 patch_prober.go:28] interesting pod/router-default-5444994796-5ftts container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 13:20:20 crc kubenswrapper[4745]: [-]has-synced failed: reason withheld Oct 10 13:20:20 crc kubenswrapper[4745]: [+]process-running ok Oct 10 13:20:20 crc kubenswrapper[4745]: healthz check failed Oct 10 13:20:20 crc kubenswrapper[4745]: I1010 13:20:20.561176 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5ftts" podUID="f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 13:20:20 crc kubenswrapper[4745]: I1010 13:20:20.969821 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5","Type":"ContainerStarted","Data":"27e389a72db7729b00092b264f13b198a7f888ad75ffd5b8065b980de23ce95c"} Oct 10 13:20:20 crc kubenswrapper[4745]: I1010 13:20:20.985804 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.985788689 podStartE2EDuration="3.985788689s" podCreationTimestamp="2025-10-10 13:20:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:20.985654455 +0000 UTC m=+134.883311228" watchObservedRunningTime="2025-10-10 13:20:20.985788689 +0000 UTC m=+134.883445452" Oct 10 13:20:21 crc kubenswrapper[4745]: I1010 13:20:21.561461 4745 patch_prober.go:28] interesting pod/router-default-5444994796-5ftts container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 13:20:21 crc kubenswrapper[4745]: [-]has-synced failed: reason withheld Oct 10 13:20:21 crc kubenswrapper[4745]: [+]process-running ok Oct 10 13:20:21 crc kubenswrapper[4745]: healthz check failed Oct 10 13:20:21 crc kubenswrapper[4745]: I1010 13:20:21.561632 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5ftts" podUID="f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 13:20:21 crc kubenswrapper[4745]: I1010 13:20:21.985826 4745 generic.go:334] "Generic (PLEG): container finished" podID="8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5" containerID="27e389a72db7729b00092b264f13b198a7f888ad75ffd5b8065b980de23ce95c" exitCode=0 Oct 10 13:20:21 crc kubenswrapper[4745]: I1010 13:20:21.985870 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5","Type":"ContainerDied","Data":"27e389a72db7729b00092b264f13b198a7f888ad75ffd5b8065b980de23ce95c"} Oct 10 13:20:22 crc kubenswrapper[4745]: I1010 13:20:22.560652 4745 patch_prober.go:28] interesting pod/router-default-5444994796-5ftts container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 13:20:22 crc kubenswrapper[4745]: [-]has-synced failed: reason withheld Oct 10 13:20:22 crc kubenswrapper[4745]: [+]process-running ok Oct 10 13:20:22 crc kubenswrapper[4745]: healthz check failed Oct 10 13:20:22 crc kubenswrapper[4745]: I1010 13:20:22.560983 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5ftts" podUID="f982b6c5-0a81-48ee-a1ca-fe2e6d10b3c9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 13:20:23 crc kubenswrapper[4745]: I1010 13:20:23.560783 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:23 crc kubenswrapper[4745]: I1010 13:20:23.566312 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-5ftts" Oct 10 13:20:25 crc kubenswrapper[4745]: I1010 13:20:25.756024 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-7jjxn" Oct 10 13:20:26 crc kubenswrapper[4745]: I1010 13:20:26.507781 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:26 crc kubenswrapper[4745]: I1010 13:20:26.512845 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:20:28 crc kubenswrapper[4745]: I1010 13:20:28.424228 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 13:20:28 crc kubenswrapper[4745]: I1010 13:20:28.527202 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5-kubelet-dir\") pod \"8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5\" (UID: \"8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5\") " Oct 10 13:20:28 crc kubenswrapper[4745]: I1010 13:20:28.527276 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5-kube-api-access\") pod \"8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5\" (UID: \"8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5\") " Oct 10 13:20:28 crc kubenswrapper[4745]: I1010 13:20:28.527334 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5" (UID: "8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:20:28 crc kubenswrapper[4745]: I1010 13:20:28.527610 4745 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 10 13:20:28 crc kubenswrapper[4745]: I1010 13:20:28.534217 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5" (UID: "8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:20:28 crc kubenswrapper[4745]: I1010 13:20:28.628567 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 13:20:29 crc kubenswrapper[4745]: I1010 13:20:29.035490 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5","Type":"ContainerDied","Data":"a7b5abd44f459c4bcfdb56b93317a67002ed6f99348a56564bf85d776944f497"} Oct 10 13:20:29 crc kubenswrapper[4745]: I1010 13:20:29.035536 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7b5abd44f459c4bcfdb56b93317a67002ed6f99348a56564bf85d776944f497" Oct 10 13:20:29 crc kubenswrapper[4745]: I1010 13:20:29.035570 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 13:20:34 crc kubenswrapper[4745]: I1010 13:20:34.159177 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:20:34 crc kubenswrapper[4745]: I1010 13:20:34.201943 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:20:34 crc kubenswrapper[4745]: I1010 13:20:34.202090 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:20:34 crc kubenswrapper[4745]: I1010 13:20:34.202168 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:20:34 crc kubenswrapper[4745]: I1010 13:20:34.203684 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:20:34 crc kubenswrapper[4745]: I1010 13:20:34.206361 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 10 13:20:34 crc kubenswrapper[4745]: I1010 13:20:34.206995 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 10 13:20:34 crc kubenswrapper[4745]: I1010 13:20:34.207237 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 10 13:20:34 crc kubenswrapper[4745]: I1010 13:20:34.216641 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 10 13:20:34 crc kubenswrapper[4745]: I1010 13:20:34.218342 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:20:34 crc kubenswrapper[4745]: I1010 13:20:34.231413 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:20:34 crc kubenswrapper[4745]: I1010 13:20:34.232031 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:20:34 crc kubenswrapper[4745]: I1010 13:20:34.257792 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:20:34 crc kubenswrapper[4745]: I1010 13:20:34.356775 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 13:20:34 crc kubenswrapper[4745]: I1010 13:20:34.459273 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:20:34 crc kubenswrapper[4745]: I1010 13:20:34.474753 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 13:20:44 crc kubenswrapper[4745]: E1010 13:20:44.091883 4745 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 10 13:20:44 crc kubenswrapper[4745]: E1010 13:20:44.092906 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h5b8x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-rp9wn_openshift-marketplace(864139db-62ff-4c5c-8478-8cf237df2f3b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 10 13:20:44 crc kubenswrapper[4745]: E1010 13:20:44.094110 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-rp9wn" podUID="864139db-62ff-4c5c-8478-8cf237df2f3b" Oct 10 13:20:45 crc kubenswrapper[4745]: E1010 13:20:45.666167 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-rp9wn" podUID="864139db-62ff-4c5c-8478-8cf237df2f3b" Oct 10 13:20:45 crc kubenswrapper[4745]: E1010 13:20:45.785868 4745 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 10 13:20:45 crc kubenswrapper[4745]: E1010 13:20:45.786042 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rrwtn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-sdzc8_openshift-marketplace(b29ff1bb-a068-485a-891d-2270ddebd191): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 10 13:20:45 crc kubenswrapper[4745]: E1010 13:20:45.788051 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-sdzc8" podUID="b29ff1bb-a068-485a-891d-2270ddebd191" Oct 10 13:20:46 crc kubenswrapper[4745]: I1010 13:20:46.187290 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:20:46 crc kubenswrapper[4745]: I1010 13:20:46.187355 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:20:46 crc kubenswrapper[4745]: I1010 13:20:46.585792 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j6gkw" Oct 10 13:20:48 crc kubenswrapper[4745]: E1010 13:20:48.542923 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-sdzc8" podUID="b29ff1bb-a068-485a-891d-2270ddebd191" Oct 10 13:20:48 crc kubenswrapper[4745]: E1010 13:20:48.620990 4745 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 10 13:20:48 crc kubenswrapper[4745]: E1010 13:20:48.621867 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ghtwr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-bbpdf_openshift-marketplace(15adca8d-0d7e-4793-a3f6-489d3aa12106): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 10 13:20:48 crc kubenswrapper[4745]: E1010 13:20:48.623524 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-bbpdf" podUID="15adca8d-0d7e-4793-a3f6-489d3aa12106" Oct 10 13:20:50 crc kubenswrapper[4745]: I1010 13:20:50.114641 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs\") pod \"network-metrics-daemon-xgh8p\" (UID: \"82c4e63a-009f-4720-87b5-640a8ec65117\") " pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:20:50 crc kubenswrapper[4745]: I1010 13:20:50.116792 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 10 13:20:50 crc kubenswrapper[4745]: I1010 13:20:50.131462 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82c4e63a-009f-4720-87b5-640a8ec65117-metrics-certs\") pod \"network-metrics-daemon-xgh8p\" (UID: \"82c4e63a-009f-4720-87b5-640a8ec65117\") " pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:20:50 crc kubenswrapper[4745]: I1010 13:20:50.369013 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 10 13:20:50 crc kubenswrapper[4745]: I1010 13:20:50.378084 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xgh8p" Oct 10 13:20:51 crc kubenswrapper[4745]: E1010 13:20:51.095523 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-bbpdf" podUID="15adca8d-0d7e-4793-a3f6-489d3aa12106" Oct 10 13:20:51 crc kubenswrapper[4745]: E1010 13:20:51.183230 4745 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 10 13:20:51 crc kubenswrapper[4745]: E1010 13:20:51.183662 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jsprc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-2k6hb_openshift-marketplace(2fa4904c-594a-4a01-aa75-e20c1ebd88af): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 10 13:20:51 crc kubenswrapper[4745]: E1010 13:20:51.185353 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-2k6hb" podUID="2fa4904c-594a-4a01-aa75-e20c1ebd88af" Oct 10 13:20:51 crc kubenswrapper[4745]: E1010 13:20:51.203877 4745 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 10 13:20:51 crc kubenswrapper[4745]: E1010 13:20:51.204072 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ff27t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-7vxfb_openshift-marketplace(0dd61fe6-4685-4ce4-9c35-6465d1aa666c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 10 13:20:51 crc kubenswrapper[4745]: E1010 13:20:51.205545 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-7vxfb" podUID="0dd61fe6-4685-4ce4-9c35-6465d1aa666c" Oct 10 13:20:51 crc kubenswrapper[4745]: E1010 13:20:51.221523 4745 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 10 13:20:51 crc kubenswrapper[4745]: E1010 13:20:51.221682 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-krclq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-tr9k9_openshift-marketplace(44dedd6d-7915-4792-81c7-d8e2dc184478): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 10 13:20:51 crc kubenswrapper[4745]: E1010 13:20:51.223074 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-tr9k9" podUID="44dedd6d-7915-4792-81c7-d8e2dc184478" Oct 10 13:20:51 crc kubenswrapper[4745]: E1010 13:20:51.248635 4745 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 10 13:20:51 crc kubenswrapper[4745]: E1010 13:20:51.248790 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ddcwd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-tb4mp_openshift-marketplace(c8674419-b98d-4f7e-ae45-b20aca88dcd5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 10 13:20:51 crc kubenswrapper[4745]: E1010 13:20:51.250158 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-tb4mp" podUID="c8674419-b98d-4f7e-ae45-b20aca88dcd5" Oct 10 13:20:51 crc kubenswrapper[4745]: W1010 13:20:51.362659 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-23e666b540c16590bd1e13b9c4ff299b70da9d85cba406433ccba1c470eb365d WatchSource:0}: Error finding container 23e666b540c16590bd1e13b9c4ff299b70da9d85cba406433ccba1c470eb365d: Status 404 returned error can't find the container with id 23e666b540c16590bd1e13b9c4ff299b70da9d85cba406433ccba1c470eb365d Oct 10 13:20:51 crc kubenswrapper[4745]: W1010 13:20:51.458870 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-29748854210ec526aa558219c800f3aec973d21f910633e1e36d31c95fe11445 WatchSource:0}: Error finding container 29748854210ec526aa558219c800f3aec973d21f910633e1e36d31c95fe11445: Status 404 returned error can't find the container with id 29748854210ec526aa558219c800f3aec973d21f910633e1e36d31c95fe11445 Oct 10 13:20:51 crc kubenswrapper[4745]: I1010 13:20:51.593426 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xgh8p"] Oct 10 13:20:51 crc kubenswrapper[4745]: W1010 13:20:51.602455 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82c4e63a_009f_4720_87b5_640a8ec65117.slice/crio-10d90dd68c7ddfd291f4b0855788316518221d37b64773bd4e7ed2549292f55a WatchSource:0}: Error finding container 10d90dd68c7ddfd291f4b0855788316518221d37b64773bd4e7ed2549292f55a: Status 404 returned error can't find the container with id 10d90dd68c7ddfd291f4b0855788316518221d37b64773bd4e7ed2549292f55a Oct 10 13:20:51 crc kubenswrapper[4745]: W1010 13:20:51.607256 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-f98c4270aa1c30f1afdcb81791faeedf79b86c38b8a602e94c29e643a3024869 WatchSource:0}: Error finding container f98c4270aa1c30f1afdcb81791faeedf79b86c38b8a602e94c29e643a3024869: Status 404 returned error can't find the container with id f98c4270aa1c30f1afdcb81791faeedf79b86c38b8a602e94c29e643a3024869 Oct 10 13:20:52 crc kubenswrapper[4745]: I1010 13:20:52.168155 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"32358f9e729f8d431097165356eae0078515ecf4bca15d33cd863e87f7b18383"} Oct 10 13:20:52 crc kubenswrapper[4745]: I1010 13:20:52.168472 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f98c4270aa1c30f1afdcb81791faeedf79b86c38b8a602e94c29e643a3024869"} Oct 10 13:20:52 crc kubenswrapper[4745]: I1010 13:20:52.168632 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:20:52 crc kubenswrapper[4745]: I1010 13:20:52.169694 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"da4ec4988c9cc2a605040f89f93430299671ef1b03db878e37439835a201708e"} Oct 10 13:20:52 crc kubenswrapper[4745]: I1010 13:20:52.169752 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"29748854210ec526aa558219c800f3aec973d21f910633e1e36d31c95fe11445"} Oct 10 13:20:52 crc kubenswrapper[4745]: I1010 13:20:52.171270 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"8540c204094e173bfcf98d5f155f6bf2abefcfab7fd52fd8032f0952a7fca8fa"} Oct 10 13:20:52 crc kubenswrapper[4745]: I1010 13:20:52.171294 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"23e666b540c16590bd1e13b9c4ff299b70da9d85cba406433ccba1c470eb365d"} Oct 10 13:20:52 crc kubenswrapper[4745]: I1010 13:20:52.173454 4745 generic.go:334] "Generic (PLEG): container finished" podID="0dbfa2cf-9426-41e9-9f46-1cfc39e2210d" containerID="28fbfe26343455ca9bc32d16a76aa4571e37028ac405e5a6e48e436a98f9a535" exitCode=0 Oct 10 13:20:52 crc kubenswrapper[4745]: I1010 13:20:52.173511 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjlxs" event={"ID":"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d","Type":"ContainerDied","Data":"28fbfe26343455ca9bc32d16a76aa4571e37028ac405e5a6e48e436a98f9a535"} Oct 10 13:20:52 crc kubenswrapper[4745]: I1010 13:20:52.175986 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" event={"ID":"82c4e63a-009f-4720-87b5-640a8ec65117","Type":"ContainerStarted","Data":"b5424ea92396cd76bebbdc147785aeb327d65694dc9b7b620b6c84b16fc17d87"} Oct 10 13:20:52 crc kubenswrapper[4745]: I1010 13:20:52.176020 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" event={"ID":"82c4e63a-009f-4720-87b5-640a8ec65117","Type":"ContainerStarted","Data":"fad56157154a972b4d7f6b224813d4af87d9d96a2c9737130e15c8f0cf7a1f0f"} Oct 10 13:20:52 crc kubenswrapper[4745]: I1010 13:20:52.176035 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xgh8p" event={"ID":"82c4e63a-009f-4720-87b5-640a8ec65117","Type":"ContainerStarted","Data":"10d90dd68c7ddfd291f4b0855788316518221d37b64773bd4e7ed2549292f55a"} Oct 10 13:20:52 crc kubenswrapper[4745]: E1010 13:20:52.177388 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-tb4mp" podUID="c8674419-b98d-4f7e-ae45-b20aca88dcd5" Oct 10 13:20:52 crc kubenswrapper[4745]: E1010 13:20:52.185071 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-tr9k9" podUID="44dedd6d-7915-4792-81c7-d8e2dc184478" Oct 10 13:20:52 crc kubenswrapper[4745]: E1010 13:20:52.185186 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-2k6hb" podUID="2fa4904c-594a-4a01-aa75-e20c1ebd88af" Oct 10 13:20:52 crc kubenswrapper[4745]: E1010 13:20:52.185276 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-7vxfb" podUID="0dd61fe6-4685-4ce4-9c35-6465d1aa666c" Oct 10 13:20:52 crc kubenswrapper[4745]: I1010 13:20:52.321010 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-xgh8p" podStartSLOduration=145.320977659 podStartE2EDuration="2m25.320977659s" podCreationTimestamp="2025-10-10 13:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:20:52.317426205 +0000 UTC m=+166.215082968" watchObservedRunningTime="2025-10-10 13:20:52.320977659 +0000 UTC m=+166.218634422" Oct 10 13:20:53 crc kubenswrapper[4745]: I1010 13:20:53.186297 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjlxs" event={"ID":"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d","Type":"ContainerStarted","Data":"ad127674c0af33fae60e0d970f31c7ad5702e9034a48dadd4a6ae254b18330fe"} Oct 10 13:20:53 crc kubenswrapper[4745]: I1010 13:20:53.200487 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mjlxs" podStartSLOduration=3.146367526 podStartE2EDuration="42.200475915s" podCreationTimestamp="2025-10-10 13:20:11 +0000 UTC" firstStartedPulling="2025-10-10 13:20:13.696024744 +0000 UTC m=+127.593681507" lastFinishedPulling="2025-10-10 13:20:52.750133133 +0000 UTC m=+166.647789896" observedRunningTime="2025-10-10 13:20:53.199341416 +0000 UTC m=+167.096998189" watchObservedRunningTime="2025-10-10 13:20:53.200475915 +0000 UTC m=+167.098132678" Oct 10 13:20:59 crc kubenswrapper[4745]: I1010 13:20:59.225604 4745 generic.go:334] "Generic (PLEG): container finished" podID="864139db-62ff-4c5c-8478-8cf237df2f3b" containerID="f3bca9758fd841780ac9cd8a0d56efb3d2d227e70c27be8651ce93b77b7c38ea" exitCode=0 Oct 10 13:20:59 crc kubenswrapper[4745]: I1010 13:20:59.225688 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rp9wn" event={"ID":"864139db-62ff-4c5c-8478-8cf237df2f3b","Type":"ContainerDied","Data":"f3bca9758fd841780ac9cd8a0d56efb3d2d227e70c27be8651ce93b77b7c38ea"} Oct 10 13:21:00 crc kubenswrapper[4745]: I1010 13:21:00.232806 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rp9wn" event={"ID":"864139db-62ff-4c5c-8478-8cf237df2f3b","Type":"ContainerStarted","Data":"1368a3156cf4245cac55530ad3ed495af9de71b8bcb1c9170dd7e55bbcc433c7"} Oct 10 13:21:00 crc kubenswrapper[4745]: I1010 13:21:00.254452 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rp9wn" podStartSLOduration=2.344814237 podStartE2EDuration="48.254426824s" podCreationTimestamp="2025-10-10 13:20:12 +0000 UTC" firstStartedPulling="2025-10-10 13:20:13.71010833 +0000 UTC m=+127.607765093" lastFinishedPulling="2025-10-10 13:20:59.619720877 +0000 UTC m=+173.517377680" observedRunningTime="2025-10-10 13:21:00.253086617 +0000 UTC m=+174.150743400" watchObservedRunningTime="2025-10-10 13:21:00.254426824 +0000 UTC m=+174.152083607" Oct 10 13:21:02 crc kubenswrapper[4745]: I1010 13:21:02.312610 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mjlxs" Oct 10 13:21:02 crc kubenswrapper[4745]: I1010 13:21:02.312980 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mjlxs" Oct 10 13:21:02 crc kubenswrapper[4745]: I1010 13:21:02.538123 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mjlxs" Oct 10 13:21:02 crc kubenswrapper[4745]: I1010 13:21:02.712972 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rp9wn" Oct 10 13:21:02 crc kubenswrapper[4745]: I1010 13:21:02.713049 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rp9wn" Oct 10 13:21:02 crc kubenswrapper[4745]: I1010 13:21:02.754284 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rp9wn" Oct 10 13:21:03 crc kubenswrapper[4745]: I1010 13:21:03.300199 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mjlxs" Oct 10 13:21:04 crc kubenswrapper[4745]: I1010 13:21:04.259646 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tr9k9" event={"ID":"44dedd6d-7915-4792-81c7-d8e2dc184478","Type":"ContainerStarted","Data":"5c805cbeec106e57e402a0564d9d743affa885c716c81c33e95c2f081d07c08a"} Oct 10 13:21:05 crc kubenswrapper[4745]: I1010 13:21:05.267248 4745 generic.go:334] "Generic (PLEG): container finished" podID="44dedd6d-7915-4792-81c7-d8e2dc184478" containerID="5c805cbeec106e57e402a0564d9d743affa885c716c81c33e95c2f081d07c08a" exitCode=0 Oct 10 13:21:05 crc kubenswrapper[4745]: I1010 13:21:05.267327 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tr9k9" event={"ID":"44dedd6d-7915-4792-81c7-d8e2dc184478","Type":"ContainerDied","Data":"5c805cbeec106e57e402a0564d9d743affa885c716c81c33e95c2f081d07c08a"} Oct 10 13:21:05 crc kubenswrapper[4745]: I1010 13:21:05.269855 4745 generic.go:334] "Generic (PLEG): container finished" podID="b29ff1bb-a068-485a-891d-2270ddebd191" containerID="8d68ecb615d5beafe50b4a1b449703802c84a884fee5ed58fdff5bb9de3854c5" exitCode=0 Oct 10 13:21:05 crc kubenswrapper[4745]: I1010 13:21:05.269978 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sdzc8" event={"ID":"b29ff1bb-a068-485a-891d-2270ddebd191","Type":"ContainerDied","Data":"8d68ecb615d5beafe50b4a1b449703802c84a884fee5ed58fdff5bb9de3854c5"} Oct 10 13:21:05 crc kubenswrapper[4745]: I1010 13:21:05.273507 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vxfb" event={"ID":"0dd61fe6-4685-4ce4-9c35-6465d1aa666c","Type":"ContainerStarted","Data":"44d1405f92c2600ab57da22329b2c006d505868286158779c7f2be57ab67886a"} Oct 10 13:21:06 crc kubenswrapper[4745]: I1010 13:21:06.282094 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sdzc8" event={"ID":"b29ff1bb-a068-485a-891d-2270ddebd191","Type":"ContainerStarted","Data":"7500e3ebcd3c5615d777ae9b538f61c8c6bf18340121a6fb8e488b9c5f583071"} Oct 10 13:21:06 crc kubenswrapper[4745]: I1010 13:21:06.285100 4745 generic.go:334] "Generic (PLEG): container finished" podID="15adca8d-0d7e-4793-a3f6-489d3aa12106" containerID="0e65b51fab8df97236a43ec0643c3fc1f9d21f0af826d31d02cb7ca1c69c0633" exitCode=0 Oct 10 13:21:06 crc kubenswrapper[4745]: I1010 13:21:06.285171 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbpdf" event={"ID":"15adca8d-0d7e-4793-a3f6-489d3aa12106","Type":"ContainerDied","Data":"0e65b51fab8df97236a43ec0643c3fc1f9d21f0af826d31d02cb7ca1c69c0633"} Oct 10 13:21:06 crc kubenswrapper[4745]: I1010 13:21:06.287485 4745 generic.go:334] "Generic (PLEG): container finished" podID="0dd61fe6-4685-4ce4-9c35-6465d1aa666c" containerID="44d1405f92c2600ab57da22329b2c006d505868286158779c7f2be57ab67886a" exitCode=0 Oct 10 13:21:06 crc kubenswrapper[4745]: I1010 13:21:06.287530 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vxfb" event={"ID":"0dd61fe6-4685-4ce4-9c35-6465d1aa666c","Type":"ContainerDied","Data":"44d1405f92c2600ab57da22329b2c006d505868286158779c7f2be57ab67886a"} Oct 10 13:21:06 crc kubenswrapper[4745]: I1010 13:21:06.295603 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tr9k9" event={"ID":"44dedd6d-7915-4792-81c7-d8e2dc184478","Type":"ContainerStarted","Data":"68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b"} Oct 10 13:21:06 crc kubenswrapper[4745]: I1010 13:21:06.311793 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sdzc8" podStartSLOduration=2.311388893 podStartE2EDuration="54.311708689s" podCreationTimestamp="2025-10-10 13:20:12 +0000 UTC" firstStartedPulling="2025-10-10 13:20:13.707403484 +0000 UTC m=+127.605060247" lastFinishedPulling="2025-10-10 13:21:05.70772327 +0000 UTC m=+179.605380043" observedRunningTime="2025-10-10 13:21:06.309083698 +0000 UTC m=+180.206740481" watchObservedRunningTime="2025-10-10 13:21:06.311708689 +0000 UTC m=+180.209365452" Oct 10 13:21:06 crc kubenswrapper[4745]: I1010 13:21:06.378530 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tr9k9" podStartSLOduration=2.066349069 podStartE2EDuration="54.378512324s" podCreationTimestamp="2025-10-10 13:20:12 +0000 UTC" firstStartedPulling="2025-10-10 13:20:13.697819724 +0000 UTC m=+127.595476487" lastFinishedPulling="2025-10-10 13:21:06.009982939 +0000 UTC m=+179.907639742" observedRunningTime="2025-10-10 13:21:06.376118691 +0000 UTC m=+180.273775454" watchObservedRunningTime="2025-10-10 13:21:06.378512324 +0000 UTC m=+180.276169087" Oct 10 13:21:07 crc kubenswrapper[4745]: I1010 13:21:07.303965 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbpdf" event={"ID":"15adca8d-0d7e-4793-a3f6-489d3aa12106","Type":"ContainerStarted","Data":"d9314152018d2b0d88c5da3474f12482068d55169b13bb77c7e2fbb5c3212689"} Oct 10 13:21:07 crc kubenswrapper[4745]: I1010 13:21:07.306538 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vxfb" event={"ID":"0dd61fe6-4685-4ce4-9c35-6465d1aa666c","Type":"ContainerStarted","Data":"a6f69285fd5a703023ac45999dd51549510b3fcb2188ac24fabc3110fa846ced"} Oct 10 13:21:07 crc kubenswrapper[4745]: I1010 13:21:07.327552 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bbpdf" podStartSLOduration=2.091967829 podStartE2EDuration="53.32753627s" podCreationTimestamp="2025-10-10 13:20:14 +0000 UTC" firstStartedPulling="2025-10-10 13:20:15.80749503 +0000 UTC m=+129.705151793" lastFinishedPulling="2025-10-10 13:21:07.043063471 +0000 UTC m=+180.940720234" observedRunningTime="2025-10-10 13:21:07.325583522 +0000 UTC m=+181.223240325" watchObservedRunningTime="2025-10-10 13:21:07.32753627 +0000 UTC m=+181.225193033" Oct 10 13:21:07 crc kubenswrapper[4745]: I1010 13:21:07.346157 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7vxfb" podStartSLOduration=2.367407489 podStartE2EDuration="52.346139167s" podCreationTimestamp="2025-10-10 13:20:15 +0000 UTC" firstStartedPulling="2025-10-10 13:20:16.823389568 +0000 UTC m=+130.721046331" lastFinishedPulling="2025-10-10 13:21:06.802121246 +0000 UTC m=+180.699778009" observedRunningTime="2025-10-10 13:21:07.345111801 +0000 UTC m=+181.242768554" watchObservedRunningTime="2025-10-10 13:21:07.346139167 +0000 UTC m=+181.243795940" Oct 10 13:21:08 crc kubenswrapper[4745]: I1010 13:21:08.314385 4745 generic.go:334] "Generic (PLEG): container finished" podID="2fa4904c-594a-4a01-aa75-e20c1ebd88af" containerID="f8bf45d714797b0eef24af51bcb209abd9ab5cd876110f5dc48259b5e829fad9" exitCode=0 Oct 10 13:21:08 crc kubenswrapper[4745]: I1010 13:21:08.314459 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2k6hb" event={"ID":"2fa4904c-594a-4a01-aa75-e20c1ebd88af","Type":"ContainerDied","Data":"f8bf45d714797b0eef24af51bcb209abd9ab5cd876110f5dc48259b5e829fad9"} Oct 10 13:21:09 crc kubenswrapper[4745]: I1010 13:21:09.322947 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2k6hb" event={"ID":"2fa4904c-594a-4a01-aa75-e20c1ebd88af","Type":"ContainerStarted","Data":"bfa30dab7fa378756782da9761380a920cb4efec952ebeac29caa0a6977b7848"} Oct 10 13:21:09 crc kubenswrapper[4745]: I1010 13:21:09.352383 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2k6hb" podStartSLOduration=2.472457601 podStartE2EDuration="54.352350454s" podCreationTimestamp="2025-10-10 13:20:15 +0000 UTC" firstStartedPulling="2025-10-10 13:20:16.838986356 +0000 UTC m=+130.736643119" lastFinishedPulling="2025-10-10 13:21:08.718879169 +0000 UTC m=+182.616535972" observedRunningTime="2025-10-10 13:21:09.343355171 +0000 UTC m=+183.241011984" watchObservedRunningTime="2025-10-10 13:21:09.352350454 +0000 UTC m=+183.250007247" Oct 10 13:21:12 crc kubenswrapper[4745]: I1010 13:21:12.779258 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rp9wn" Oct 10 13:21:12 crc kubenswrapper[4745]: I1010 13:21:12.790700 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tr9k9" Oct 10 13:21:12 crc kubenswrapper[4745]: I1010 13:21:12.790768 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tr9k9" Oct 10 13:21:12 crc kubenswrapper[4745]: I1010 13:21:12.842158 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tr9k9" Oct 10 13:21:12 crc kubenswrapper[4745]: I1010 13:21:12.936545 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sdzc8" Oct 10 13:21:12 crc kubenswrapper[4745]: I1010 13:21:12.936606 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sdzc8" Oct 10 13:21:13 crc kubenswrapper[4745]: I1010 13:21:13.087544 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sdzc8" Oct 10 13:21:13 crc kubenswrapper[4745]: I1010 13:21:13.406072 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tr9k9" Oct 10 13:21:13 crc kubenswrapper[4745]: I1010 13:21:13.412966 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sdzc8" Oct 10 13:21:14 crc kubenswrapper[4745]: I1010 13:21:14.351539 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tb4mp" event={"ID":"c8674419-b98d-4f7e-ae45-b20aca88dcd5","Type":"ContainerStarted","Data":"6ff7678a4d663ac651efab7ce6c01c82a34c0d1ece1e6cb83886015454c984e6"} Oct 10 13:21:14 crc kubenswrapper[4745]: I1010 13:21:14.568882 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sdzc8"] Oct 10 13:21:14 crc kubenswrapper[4745]: I1010 13:21:14.915661 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bbpdf" Oct 10 13:21:14 crc kubenswrapper[4745]: I1010 13:21:14.915760 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bbpdf" Oct 10 13:21:14 crc kubenswrapper[4745]: I1010 13:21:14.979011 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bbpdf" Oct 10 13:21:15 crc kubenswrapper[4745]: I1010 13:21:15.169434 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rp9wn"] Oct 10 13:21:15 crc kubenswrapper[4745]: I1010 13:21:15.169874 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rp9wn" podUID="864139db-62ff-4c5c-8478-8cf237df2f3b" containerName="registry-server" containerID="cri-o://1368a3156cf4245cac55530ad3ed495af9de71b8bcb1c9170dd7e55bbcc433c7" gracePeriod=2 Oct 10 13:21:15 crc kubenswrapper[4745]: I1010 13:21:15.360516 4745 generic.go:334] "Generic (PLEG): container finished" podID="c8674419-b98d-4f7e-ae45-b20aca88dcd5" containerID="6ff7678a4d663ac651efab7ce6c01c82a34c0d1ece1e6cb83886015454c984e6" exitCode=0 Oct 10 13:21:15 crc kubenswrapper[4745]: I1010 13:21:15.360582 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tb4mp" event={"ID":"c8674419-b98d-4f7e-ae45-b20aca88dcd5","Type":"ContainerDied","Data":"6ff7678a4d663ac651efab7ce6c01c82a34c0d1ece1e6cb83886015454c984e6"} Oct 10 13:21:15 crc kubenswrapper[4745]: I1010 13:21:15.405543 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bbpdf" Oct 10 13:21:15 crc kubenswrapper[4745]: I1010 13:21:15.489376 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7vxfb" Oct 10 13:21:15 crc kubenswrapper[4745]: I1010 13:21:15.489463 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7vxfb" Oct 10 13:21:15 crc kubenswrapper[4745]: I1010 13:21:15.525324 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7vxfb" Oct 10 13:21:15 crc kubenswrapper[4745]: I1010 13:21:15.893665 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2k6hb" Oct 10 13:21:15 crc kubenswrapper[4745]: I1010 13:21:15.894065 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2k6hb" Oct 10 13:21:15 crc kubenswrapper[4745]: I1010 13:21:15.956658 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2k6hb" Oct 10 13:21:16 crc kubenswrapper[4745]: I1010 13:21:16.187227 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:21:16 crc kubenswrapper[4745]: I1010 13:21:16.187293 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:21:16 crc kubenswrapper[4745]: I1010 13:21:16.370599 4745 generic.go:334] "Generic (PLEG): container finished" podID="864139db-62ff-4c5c-8478-8cf237df2f3b" containerID="1368a3156cf4245cac55530ad3ed495af9de71b8bcb1c9170dd7e55bbcc433c7" exitCode=0 Oct 10 13:21:16 crc kubenswrapper[4745]: I1010 13:21:16.371172 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sdzc8" podUID="b29ff1bb-a068-485a-891d-2270ddebd191" containerName="registry-server" containerID="cri-o://7500e3ebcd3c5615d777ae9b538f61c8c6bf18340121a6fb8e488b9c5f583071" gracePeriod=2 Oct 10 13:21:16 crc kubenswrapper[4745]: I1010 13:21:16.371716 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rp9wn" event={"ID":"864139db-62ff-4c5c-8478-8cf237df2f3b","Type":"ContainerDied","Data":"1368a3156cf4245cac55530ad3ed495af9de71b8bcb1c9170dd7e55bbcc433c7"} Oct 10 13:21:16 crc kubenswrapper[4745]: I1010 13:21:16.411423 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7vxfb" Oct 10 13:21:16 crc kubenswrapper[4745]: I1010 13:21:16.426115 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2k6hb" Oct 10 13:21:17 crc kubenswrapper[4745]: I1010 13:21:17.573293 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bbpdf"] Oct 10 13:21:17 crc kubenswrapper[4745]: I1010 13:21:17.573645 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bbpdf" podUID="15adca8d-0d7e-4793-a3f6-489d3aa12106" containerName="registry-server" containerID="cri-o://d9314152018d2b0d88c5da3474f12482068d55169b13bb77c7e2fbb5c3212689" gracePeriod=2 Oct 10 13:21:17 crc kubenswrapper[4745]: I1010 13:21:17.753854 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rp9wn" Oct 10 13:21:17 crc kubenswrapper[4745]: I1010 13:21:17.892151 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/864139db-62ff-4c5c-8478-8cf237df2f3b-catalog-content\") pod \"864139db-62ff-4c5c-8478-8cf237df2f3b\" (UID: \"864139db-62ff-4c5c-8478-8cf237df2f3b\") " Oct 10 13:21:17 crc kubenswrapper[4745]: I1010 13:21:17.892227 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5b8x\" (UniqueName: \"kubernetes.io/projected/864139db-62ff-4c5c-8478-8cf237df2f3b-kube-api-access-h5b8x\") pod \"864139db-62ff-4c5c-8478-8cf237df2f3b\" (UID: \"864139db-62ff-4c5c-8478-8cf237df2f3b\") " Oct 10 13:21:17 crc kubenswrapper[4745]: I1010 13:21:17.892287 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/864139db-62ff-4c5c-8478-8cf237df2f3b-utilities\") pod \"864139db-62ff-4c5c-8478-8cf237df2f3b\" (UID: \"864139db-62ff-4c5c-8478-8cf237df2f3b\") " Oct 10 13:21:17 crc kubenswrapper[4745]: I1010 13:21:17.894101 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/864139db-62ff-4c5c-8478-8cf237df2f3b-utilities" (OuterVolumeSpecName: "utilities") pod "864139db-62ff-4c5c-8478-8cf237df2f3b" (UID: "864139db-62ff-4c5c-8478-8cf237df2f3b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:21:17 crc kubenswrapper[4745]: I1010 13:21:17.900382 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/864139db-62ff-4c5c-8478-8cf237df2f3b-kube-api-access-h5b8x" (OuterVolumeSpecName: "kube-api-access-h5b8x") pod "864139db-62ff-4c5c-8478-8cf237df2f3b" (UID: "864139db-62ff-4c5c-8478-8cf237df2f3b"). InnerVolumeSpecName "kube-api-access-h5b8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:21:17 crc kubenswrapper[4745]: I1010 13:21:17.935610 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/864139db-62ff-4c5c-8478-8cf237df2f3b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "864139db-62ff-4c5c-8478-8cf237df2f3b" (UID: "864139db-62ff-4c5c-8478-8cf237df2f3b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:21:17 crc kubenswrapper[4745]: I1010 13:21:17.994637 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/864139db-62ff-4c5c-8478-8cf237df2f3b-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:17 crc kubenswrapper[4745]: I1010 13:21:17.994711 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/864139db-62ff-4c5c-8478-8cf237df2f3b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:17 crc kubenswrapper[4745]: I1010 13:21:17.994778 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5b8x\" (UniqueName: \"kubernetes.io/projected/864139db-62ff-4c5c-8478-8cf237df2f3b-kube-api-access-h5b8x\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:18 crc kubenswrapper[4745]: I1010 13:21:18.390604 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rp9wn" event={"ID":"864139db-62ff-4c5c-8478-8cf237df2f3b","Type":"ContainerDied","Data":"ead87654d18f0dd9bc336113704ca06da965ec8fff7206a5fe94c8ad805bbdcd"} Oct 10 13:21:18 crc kubenswrapper[4745]: I1010 13:21:18.390658 4745 scope.go:117] "RemoveContainer" containerID="1368a3156cf4245cac55530ad3ed495af9de71b8bcb1c9170dd7e55bbcc433c7" Oct 10 13:21:18 crc kubenswrapper[4745]: I1010 13:21:18.390815 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rp9wn" Oct 10 13:21:18 crc kubenswrapper[4745]: I1010 13:21:18.438295 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rp9wn"] Oct 10 13:21:18 crc kubenswrapper[4745]: I1010 13:21:18.441050 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rp9wn"] Oct 10 13:21:18 crc kubenswrapper[4745]: I1010 13:21:18.480288 4745 scope.go:117] "RemoveContainer" containerID="f3bca9758fd841780ac9cd8a0d56efb3d2d227e70c27be8651ce93b77b7c38ea" Oct 10 13:21:18 crc kubenswrapper[4745]: I1010 13:21:18.500602 4745 scope.go:117] "RemoveContainer" containerID="7a5f4460de68703a5b6619368deeb9bb92506dcebbf0d9c0d3d92488bdc76531" Oct 10 13:21:18 crc kubenswrapper[4745]: I1010 13:21:18.751443 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="864139db-62ff-4c5c-8478-8cf237df2f3b" path="/var/lib/kubelet/pods/864139db-62ff-4c5c-8478-8cf237df2f3b/volumes" Oct 10 13:21:19 crc kubenswrapper[4745]: I1010 13:21:19.398997 4745 generic.go:334] "Generic (PLEG): container finished" podID="b29ff1bb-a068-485a-891d-2270ddebd191" containerID="7500e3ebcd3c5615d777ae9b538f61c8c6bf18340121a6fb8e488b9c5f583071" exitCode=0 Oct 10 13:21:19 crc kubenswrapper[4745]: I1010 13:21:19.399099 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sdzc8" event={"ID":"b29ff1bb-a068-485a-891d-2270ddebd191","Type":"ContainerDied","Data":"7500e3ebcd3c5615d777ae9b538f61c8c6bf18340121a6fb8e488b9c5f583071"} Oct 10 13:21:19 crc kubenswrapper[4745]: I1010 13:21:19.971834 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2k6hb"] Oct 10 13:21:19 crc kubenswrapper[4745]: I1010 13:21:19.972784 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2k6hb" podUID="2fa4904c-594a-4a01-aa75-e20c1ebd88af" containerName="registry-server" containerID="cri-o://bfa30dab7fa378756782da9761380a920cb4efec952ebeac29caa0a6977b7848" gracePeriod=2 Oct 10 13:21:20 crc kubenswrapper[4745]: I1010 13:21:20.409295 4745 generic.go:334] "Generic (PLEG): container finished" podID="15adca8d-0d7e-4793-a3f6-489d3aa12106" containerID="d9314152018d2b0d88c5da3474f12482068d55169b13bb77c7e2fbb5c3212689" exitCode=0 Oct 10 13:21:20 crc kubenswrapper[4745]: I1010 13:21:20.409393 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbpdf" event={"ID":"15adca8d-0d7e-4793-a3f6-489d3aa12106","Type":"ContainerDied","Data":"d9314152018d2b0d88c5da3474f12482068d55169b13bb77c7e2fbb5c3212689"} Oct 10 13:21:20 crc kubenswrapper[4745]: I1010 13:21:20.928173 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sdzc8" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.032524 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b29ff1bb-a068-485a-891d-2270ddebd191-catalog-content\") pod \"b29ff1bb-a068-485a-891d-2270ddebd191\" (UID: \"b29ff1bb-a068-485a-891d-2270ddebd191\") " Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.032605 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrwtn\" (UniqueName: \"kubernetes.io/projected/b29ff1bb-a068-485a-891d-2270ddebd191-kube-api-access-rrwtn\") pod \"b29ff1bb-a068-485a-891d-2270ddebd191\" (UID: \"b29ff1bb-a068-485a-891d-2270ddebd191\") " Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.032700 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b29ff1bb-a068-485a-891d-2270ddebd191-utilities\") pod \"b29ff1bb-a068-485a-891d-2270ddebd191\" (UID: \"b29ff1bb-a068-485a-891d-2270ddebd191\") " Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.034059 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b29ff1bb-a068-485a-891d-2270ddebd191-utilities" (OuterVolumeSpecName: "utilities") pod "b29ff1bb-a068-485a-891d-2270ddebd191" (UID: "b29ff1bb-a068-485a-891d-2270ddebd191"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.040001 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b29ff1bb-a068-485a-891d-2270ddebd191-kube-api-access-rrwtn" (OuterVolumeSpecName: "kube-api-access-rrwtn") pod "b29ff1bb-a068-485a-891d-2270ddebd191" (UID: "b29ff1bb-a068-485a-891d-2270ddebd191"). InnerVolumeSpecName "kube-api-access-rrwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.080511 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b29ff1bb-a068-485a-891d-2270ddebd191-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b29ff1bb-a068-485a-891d-2270ddebd191" (UID: "b29ff1bb-a068-485a-891d-2270ddebd191"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.134506 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b29ff1bb-a068-485a-891d-2270ddebd191-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.134593 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrwtn\" (UniqueName: \"kubernetes.io/projected/b29ff1bb-a068-485a-891d-2270ddebd191-kube-api-access-rrwtn\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.134615 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b29ff1bb-a068-485a-891d-2270ddebd191-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.166779 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bbpdf" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.337278 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15adca8d-0d7e-4793-a3f6-489d3aa12106-utilities\") pod \"15adca8d-0d7e-4793-a3f6-489d3aa12106\" (UID: \"15adca8d-0d7e-4793-a3f6-489d3aa12106\") " Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.337429 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15adca8d-0d7e-4793-a3f6-489d3aa12106-catalog-content\") pod \"15adca8d-0d7e-4793-a3f6-489d3aa12106\" (UID: \"15adca8d-0d7e-4793-a3f6-489d3aa12106\") " Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.337485 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghtwr\" (UniqueName: \"kubernetes.io/projected/15adca8d-0d7e-4793-a3f6-489d3aa12106-kube-api-access-ghtwr\") pod \"15adca8d-0d7e-4793-a3f6-489d3aa12106\" (UID: \"15adca8d-0d7e-4793-a3f6-489d3aa12106\") " Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.338119 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15adca8d-0d7e-4793-a3f6-489d3aa12106-utilities" (OuterVolumeSpecName: "utilities") pod "15adca8d-0d7e-4793-a3f6-489d3aa12106" (UID: "15adca8d-0d7e-4793-a3f6-489d3aa12106"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.341150 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15adca8d-0d7e-4793-a3f6-489d3aa12106-kube-api-access-ghtwr" (OuterVolumeSpecName: "kube-api-access-ghtwr") pod "15adca8d-0d7e-4793-a3f6-489d3aa12106" (UID: "15adca8d-0d7e-4793-a3f6-489d3aa12106"). InnerVolumeSpecName "kube-api-access-ghtwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.351815 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15adca8d-0d7e-4793-a3f6-489d3aa12106-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15adca8d-0d7e-4793-a3f6-489d3aa12106" (UID: "15adca8d-0d7e-4793-a3f6-489d3aa12106"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.417701 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbpdf" event={"ID":"15adca8d-0d7e-4793-a3f6-489d3aa12106","Type":"ContainerDied","Data":"e42ce376fcacc2b2ee434f439139cfb9cb33cd002e0c53212777d701a251194b"} Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.417784 4745 scope.go:117] "RemoveContainer" containerID="d9314152018d2b0d88c5da3474f12482068d55169b13bb77c7e2fbb5c3212689" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.417792 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bbpdf" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.422170 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sdzc8" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.424031 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sdzc8" event={"ID":"b29ff1bb-a068-485a-891d-2270ddebd191","Type":"ContainerDied","Data":"c76ef6884d858922712c8592c0fcf6729f0c738bd260253f63022bb5572594f6"} Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.427571 4745 generic.go:334] "Generic (PLEG): container finished" podID="2fa4904c-594a-4a01-aa75-e20c1ebd88af" containerID="bfa30dab7fa378756782da9761380a920cb4efec952ebeac29caa0a6977b7848" exitCode=0 Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.427638 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2k6hb" event={"ID":"2fa4904c-594a-4a01-aa75-e20c1ebd88af","Type":"ContainerDied","Data":"bfa30dab7fa378756782da9761380a920cb4efec952ebeac29caa0a6977b7848"} Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.438764 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15adca8d-0d7e-4793-a3f6-489d3aa12106-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.439163 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15adca8d-0d7e-4793-a3f6-489d3aa12106-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.439364 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghtwr\" (UniqueName: \"kubernetes.io/projected/15adca8d-0d7e-4793-a3f6-489d3aa12106-kube-api-access-ghtwr\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.464343 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sdzc8"] Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.475695 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sdzc8"] Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.479174 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bbpdf"] Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.482313 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bbpdf"] Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.939574 4745 scope.go:117] "RemoveContainer" containerID="0e65b51fab8df97236a43ec0643c3fc1f9d21f0af826d31d02cb7ca1c69c0633" Oct 10 13:21:21 crc kubenswrapper[4745]: I1010 13:21:21.984106 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2k6hb" Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.149763 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsprc\" (UniqueName: \"kubernetes.io/projected/2fa4904c-594a-4a01-aa75-e20c1ebd88af-kube-api-access-jsprc\") pod \"2fa4904c-594a-4a01-aa75-e20c1ebd88af\" (UID: \"2fa4904c-594a-4a01-aa75-e20c1ebd88af\") " Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.149919 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fa4904c-594a-4a01-aa75-e20c1ebd88af-catalog-content\") pod \"2fa4904c-594a-4a01-aa75-e20c1ebd88af\" (UID: \"2fa4904c-594a-4a01-aa75-e20c1ebd88af\") " Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.150010 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fa4904c-594a-4a01-aa75-e20c1ebd88af-utilities\") pod \"2fa4904c-594a-4a01-aa75-e20c1ebd88af\" (UID: \"2fa4904c-594a-4a01-aa75-e20c1ebd88af\") " Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.151782 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fa4904c-594a-4a01-aa75-e20c1ebd88af-utilities" (OuterVolumeSpecName: "utilities") pod "2fa4904c-594a-4a01-aa75-e20c1ebd88af" (UID: "2fa4904c-594a-4a01-aa75-e20c1ebd88af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.154505 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fa4904c-594a-4a01-aa75-e20c1ebd88af-kube-api-access-jsprc" (OuterVolumeSpecName: "kube-api-access-jsprc") pod "2fa4904c-594a-4a01-aa75-e20c1ebd88af" (UID: "2fa4904c-594a-4a01-aa75-e20c1ebd88af"). InnerVolumeSpecName "kube-api-access-jsprc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.262984 4745 scope.go:117] "RemoveContainer" containerID="191efae0e5d2efb79650d7a64835385de626bd322644f6646e08d3c890ee94f0" Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.263715 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fa4904c-594a-4a01-aa75-e20c1ebd88af-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.263772 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsprc\" (UniqueName: \"kubernetes.io/projected/2fa4904c-594a-4a01-aa75-e20c1ebd88af-kube-api-access-jsprc\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.270201 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fa4904c-594a-4a01-aa75-e20c1ebd88af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2fa4904c-594a-4a01-aa75-e20c1ebd88af" (UID: "2fa4904c-594a-4a01-aa75-e20c1ebd88af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.285133 4745 scope.go:117] "RemoveContainer" containerID="7500e3ebcd3c5615d777ae9b538f61c8c6bf18340121a6fb8e488b9c5f583071" Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.300641 4745 scope.go:117] "RemoveContainer" containerID="8d68ecb615d5beafe50b4a1b449703802c84a884fee5ed58fdff5bb9de3854c5" Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.345071 4745 scope.go:117] "RemoveContainer" containerID="46f975de84b406276106bc9e1135fea1ff94946bd0882d033fc240490b1a7dd8" Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.365062 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fa4904c-594a-4a01-aa75-e20c1ebd88af-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.451514 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2k6hb" event={"ID":"2fa4904c-594a-4a01-aa75-e20c1ebd88af","Type":"ContainerDied","Data":"bbb094c761ad0989c623cd7a86ed17431b16dcdde290dfbd8eaed91aea116523"} Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.451541 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2k6hb" Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.451578 4745 scope.go:117] "RemoveContainer" containerID="bfa30dab7fa378756782da9761380a920cb4efec952ebeac29caa0a6977b7848" Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.471558 4745 scope.go:117] "RemoveContainer" containerID="f8bf45d714797b0eef24af51bcb209abd9ab5cd876110f5dc48259b5e829fad9" Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.489695 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2k6hb"] Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.499594 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2k6hb"] Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.513442 4745 scope.go:117] "RemoveContainer" containerID="decb9e8b0ad4f92623615f0efd7ea7790efac205503de47088372c1ce3318556" Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.752144 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15adca8d-0d7e-4793-a3f6-489d3aa12106" path="/var/lib/kubelet/pods/15adca8d-0d7e-4793-a3f6-489d3aa12106/volumes" Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.753760 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fa4904c-594a-4a01-aa75-e20c1ebd88af" path="/var/lib/kubelet/pods/2fa4904c-594a-4a01-aa75-e20c1ebd88af/volumes" Oct 10 13:21:22 crc kubenswrapper[4745]: I1010 13:21:22.754976 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b29ff1bb-a068-485a-891d-2270ddebd191" path="/var/lib/kubelet/pods/b29ff1bb-a068-485a-891d-2270ddebd191/volumes" Oct 10 13:21:23 crc kubenswrapper[4745]: I1010 13:21:23.460385 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tb4mp" event={"ID":"c8674419-b98d-4f7e-ae45-b20aca88dcd5","Type":"ContainerStarted","Data":"0576a9b692231d5a65ac55573a29dd12d44a79a12bd786d50f7c6a29aeb181e1"} Oct 10 13:21:23 crc kubenswrapper[4745]: I1010 13:21:23.477720 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tb4mp" podStartSLOduration=3.016975853 podStartE2EDuration="1m9.477705735s" podCreationTimestamp="2025-10-10 13:20:14 +0000 UTC" firstStartedPulling="2025-10-10 13:20:15.803440026 +0000 UTC m=+129.701096789" lastFinishedPulling="2025-10-10 13:21:22.264169898 +0000 UTC m=+196.161826671" observedRunningTime="2025-10-10 13:21:23.474247997 +0000 UTC m=+197.371904760" watchObservedRunningTime="2025-10-10 13:21:23.477705735 +0000 UTC m=+197.375362498" Oct 10 13:21:24 crc kubenswrapper[4745]: I1010 13:21:24.472302 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 13:21:24 crc kubenswrapper[4745]: I1010 13:21:24.502490 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tb4mp" Oct 10 13:21:24 crc kubenswrapper[4745]: I1010 13:21:24.502556 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tb4mp" Oct 10 13:21:24 crc kubenswrapper[4745]: I1010 13:21:24.563430 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tb4mp" Oct 10 13:21:24 crc kubenswrapper[4745]: I1010 13:21:24.730767 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-98dn4"] Oct 10 13:21:34 crc kubenswrapper[4745]: I1010 13:21:34.542971 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tb4mp" Oct 10 13:21:46 crc kubenswrapper[4745]: I1010 13:21:46.187199 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:21:46 crc kubenswrapper[4745]: I1010 13:21:46.187756 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:21:46 crc kubenswrapper[4745]: I1010 13:21:46.187811 4745 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:21:46 crc kubenswrapper[4745]: I1010 13:21:46.188236 4745 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb"} pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 13:21:46 crc kubenswrapper[4745]: I1010 13:21:46.188331 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" containerID="cri-o://01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb" gracePeriod=600 Oct 10 13:21:46 crc kubenswrapper[4745]: I1010 13:21:46.579711 4745 generic.go:334] "Generic (PLEG): container finished" podID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerID="01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb" exitCode=0 Oct 10 13:21:46 crc kubenswrapper[4745]: I1010 13:21:46.579829 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerDied","Data":"01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb"} Oct 10 13:21:46 crc kubenswrapper[4745]: I1010 13:21:46.579873 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"60a361e1a45479d481d0afd09183405e3656ee9875dea13f2df640bcec7617c6"} Oct 10 13:21:49 crc kubenswrapper[4745]: I1010 13:21:49.757239 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" podUID="11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" containerName="oauth-openshift" containerID="cri-o://539d549721a124f489450d0d93f931208c97452ef2c7fad3f880aee6231ba8da" gracePeriod=15 Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.168915 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.202756 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-57864598c6-v5hh6"] Oct 10 13:21:50 crc kubenswrapper[4745]: E1010 13:21:50.203016 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="864139db-62ff-4c5c-8478-8cf237df2f3b" containerName="extract-utilities" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203039 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="864139db-62ff-4c5c-8478-8cf237df2f3b" containerName="extract-utilities" Oct 10 13:21:50 crc kubenswrapper[4745]: E1010 13:21:50.203061 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b29ff1bb-a068-485a-891d-2270ddebd191" containerName="extract-utilities" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203069 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="b29ff1bb-a068-485a-891d-2270ddebd191" containerName="extract-utilities" Oct 10 13:21:50 crc kubenswrapper[4745]: E1010 13:21:50.203082 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15adca8d-0d7e-4793-a3f6-489d3aa12106" containerName="extract-content" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203089 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="15adca8d-0d7e-4793-a3f6-489d3aa12106" containerName="extract-content" Oct 10 13:21:50 crc kubenswrapper[4745]: E1010 13:21:50.203101 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fa4904c-594a-4a01-aa75-e20c1ebd88af" containerName="registry-server" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203109 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fa4904c-594a-4a01-aa75-e20c1ebd88af" containerName="registry-server" Oct 10 13:21:50 crc kubenswrapper[4745]: E1010 13:21:50.203118 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" containerName="oauth-openshift" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203126 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" containerName="oauth-openshift" Oct 10 13:21:50 crc kubenswrapper[4745]: E1010 13:21:50.203136 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb8607d-1cbd-4b3a-9b06-f2db94952d82" containerName="collect-profiles" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203144 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb8607d-1cbd-4b3a-9b06-f2db94952d82" containerName="collect-profiles" Oct 10 13:21:50 crc kubenswrapper[4745]: E1010 13:21:50.203156 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5" containerName="pruner" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203164 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5" containerName="pruner" Oct 10 13:21:50 crc kubenswrapper[4745]: E1010 13:21:50.203173 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fa4904c-594a-4a01-aa75-e20c1ebd88af" containerName="extract-content" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203181 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fa4904c-594a-4a01-aa75-e20c1ebd88af" containerName="extract-content" Oct 10 13:21:50 crc kubenswrapper[4745]: E1010 13:21:50.203190 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fa4904c-594a-4a01-aa75-e20c1ebd88af" containerName="extract-utilities" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203197 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fa4904c-594a-4a01-aa75-e20c1ebd88af" containerName="extract-utilities" Oct 10 13:21:50 crc kubenswrapper[4745]: E1010 13:21:50.203209 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="864139db-62ff-4c5c-8478-8cf237df2f3b" containerName="registry-server" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203216 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="864139db-62ff-4c5c-8478-8cf237df2f3b" containerName="registry-server" Oct 10 13:21:50 crc kubenswrapper[4745]: E1010 13:21:50.203227 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="864139db-62ff-4c5c-8478-8cf237df2f3b" containerName="extract-content" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203234 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="864139db-62ff-4c5c-8478-8cf237df2f3b" containerName="extract-content" Oct 10 13:21:50 crc kubenswrapper[4745]: E1010 13:21:50.203246 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b29ff1bb-a068-485a-891d-2270ddebd191" containerName="registry-server" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203253 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="b29ff1bb-a068-485a-891d-2270ddebd191" containerName="registry-server" Oct 10 13:21:50 crc kubenswrapper[4745]: E1010 13:21:50.203265 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b29ff1bb-a068-485a-891d-2270ddebd191" containerName="extract-content" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203272 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="b29ff1bb-a068-485a-891d-2270ddebd191" containerName="extract-content" Oct 10 13:21:50 crc kubenswrapper[4745]: E1010 13:21:50.203280 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15adca8d-0d7e-4793-a3f6-489d3aa12106" containerName="registry-server" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203288 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="15adca8d-0d7e-4793-a3f6-489d3aa12106" containerName="registry-server" Oct 10 13:21:50 crc kubenswrapper[4745]: E1010 13:21:50.203299 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15adca8d-0d7e-4793-a3f6-489d3aa12106" containerName="extract-utilities" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203306 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="15adca8d-0d7e-4793-a3f6-489d3aa12106" containerName="extract-utilities" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203418 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="8051b8a9-7a9c-4fb8-83d7-8067fd8ab4d5" containerName="pruner" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203433 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" containerName="oauth-openshift" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203446 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cb8607d-1cbd-4b3a-9b06-f2db94952d82" containerName="collect-profiles" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203463 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="b29ff1bb-a068-485a-891d-2270ddebd191" containerName="registry-server" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203475 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="15adca8d-0d7e-4793-a3f6-489d3aa12106" containerName="registry-server" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203489 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="864139db-62ff-4c5c-8478-8cf237df2f3b" containerName="registry-server" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.203501 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fa4904c-594a-4a01-aa75-e20c1ebd88af" containerName="registry-server" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.204068 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.224564 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57864598c6-v5hh6"] Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228065 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-ocp-branding-template\") pod \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228171 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-session\") pod \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228208 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-audit-policies\") pod \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228340 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-error\") pod \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228414 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-trusted-ca-bundle\") pod \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228438 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-serving-cert\") pod \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228462 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-router-certs\") pod \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228503 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-service-ca\") pod \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228533 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-cliconfig\") pod \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228553 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk4gk\" (UniqueName: \"kubernetes.io/projected/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-kube-api-access-tk4gk\") pod \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228577 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-audit-dir\") pod \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228600 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-idp-0-file-data\") pod \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228623 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-provider-selection\") pod \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228649 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-login\") pod \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\" (UID: \"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3\") " Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228866 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-user-template-error\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228912 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228937 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qbsz\" (UniqueName: \"kubernetes.io/projected/238cafef-e181-4af4-9e19-b5b5e3db5fd8-kube-api-access-4qbsz\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.228962 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.229001 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/238cafef-e181-4af4-9e19-b5b5e3db5fd8-audit-dir\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.229020 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.229041 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-user-template-login\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.229068 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-service-ca\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.229109 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-router-certs\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.229133 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-session\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.229156 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.229184 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.229210 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.229234 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/238cafef-e181-4af4-9e19-b5b5e3db5fd8-audit-policies\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.229886 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" (UID: "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.229970 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" (UID: "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.233292 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" (UID: "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.233470 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" (UID: "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.233820 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" (UID: "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.236639 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" (UID: "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.237011 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" (UID: "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.237591 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-kube-api-access-tk4gk" (OuterVolumeSpecName: "kube-api-access-tk4gk") pod "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" (UID: "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3"). InnerVolumeSpecName "kube-api-access-tk4gk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.251286 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" (UID: "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.251484 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" (UID: "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.251557 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" (UID: "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.251769 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" (UID: "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.252797 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" (UID: "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.253217 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" (UID: "11d3c997-e4ad-49fc-9d99-8a646f0fc5c3"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.330691 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.330761 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.330786 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/238cafef-e181-4af4-9e19-b5b5e3db5fd8-audit-policies\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.330837 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-user-template-error\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.330868 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.330885 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qbsz\" (UniqueName: \"kubernetes.io/projected/238cafef-e181-4af4-9e19-b5b5e3db5fd8-kube-api-access-4qbsz\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.330905 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.330933 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/238cafef-e181-4af4-9e19-b5b5e3db5fd8-audit-dir\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.330948 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.330965 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-user-template-login\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.330987 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-service-ca\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331006 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-router-certs\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331022 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-session\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331039 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331074 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331085 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331097 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331106 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331115 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331124 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk4gk\" (UniqueName: \"kubernetes.io/projected/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-kube-api-access-tk4gk\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331134 4745 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331142 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331152 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331161 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331169 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331180 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331188 4745 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331196 4745 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.331579 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.332519 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/238cafef-e181-4af4-9e19-b5b5e3db5fd8-audit-dir\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.332689 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.333636 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-service-ca\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.333719 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/238cafef-e181-4af4-9e19-b5b5e3db5fd8-audit-policies\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.334693 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-user-template-error\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.334746 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.335127 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.336178 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-router-certs\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.336549 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-user-template-login\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.336721 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.337106 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-system-session\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.338033 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/238cafef-e181-4af4-9e19-b5b5e3db5fd8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.350127 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qbsz\" (UniqueName: \"kubernetes.io/projected/238cafef-e181-4af4-9e19-b5b5e3db5fd8-kube-api-access-4qbsz\") pod \"oauth-openshift-57864598c6-v5hh6\" (UID: \"238cafef-e181-4af4-9e19-b5b5e3db5fd8\") " pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.580271 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.605353 4745 generic.go:334] "Generic (PLEG): container finished" podID="11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" containerID="539d549721a124f489450d0d93f931208c97452ef2c7fad3f880aee6231ba8da" exitCode=0 Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.605411 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" event={"ID":"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3","Type":"ContainerDied","Data":"539d549721a124f489450d0d93f931208c97452ef2c7fad3f880aee6231ba8da"} Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.605421 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.605458 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-98dn4" event={"ID":"11d3c997-e4ad-49fc-9d99-8a646f0fc5c3","Type":"ContainerDied","Data":"dc9a1d275d3a25763aeb84005a5a2a0687c95c3ab2289f74082126002bea01f7"} Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.605485 4745 scope.go:117] "RemoveContainer" containerID="539d549721a124f489450d0d93f931208c97452ef2c7fad3f880aee6231ba8da" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.651981 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-98dn4"] Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.658203 4745 scope.go:117] "RemoveContainer" containerID="539d549721a124f489450d0d93f931208c97452ef2c7fad3f880aee6231ba8da" Oct 10 13:21:50 crc kubenswrapper[4745]: E1010 13:21:50.659362 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"539d549721a124f489450d0d93f931208c97452ef2c7fad3f880aee6231ba8da\": container with ID starting with 539d549721a124f489450d0d93f931208c97452ef2c7fad3f880aee6231ba8da not found: ID does not exist" containerID="539d549721a124f489450d0d93f931208c97452ef2c7fad3f880aee6231ba8da" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.659450 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"539d549721a124f489450d0d93f931208c97452ef2c7fad3f880aee6231ba8da"} err="failed to get container status \"539d549721a124f489450d0d93f931208c97452ef2c7fad3f880aee6231ba8da\": rpc error: code = NotFound desc = could not find container \"539d549721a124f489450d0d93f931208c97452ef2c7fad3f880aee6231ba8da\": container with ID starting with 539d549721a124f489450d0d93f931208c97452ef2c7fad3f880aee6231ba8da not found: ID does not exist" Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.660464 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-98dn4"] Oct 10 13:21:50 crc kubenswrapper[4745]: I1010 13:21:50.756949 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11d3c997-e4ad-49fc-9d99-8a646f0fc5c3" path="/var/lib/kubelet/pods/11d3c997-e4ad-49fc-9d99-8a646f0fc5c3/volumes" Oct 10 13:21:51 crc kubenswrapper[4745]: I1010 13:21:51.061803 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57864598c6-v5hh6"] Oct 10 13:21:51 crc kubenswrapper[4745]: I1010 13:21:51.610954 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" event={"ID":"238cafef-e181-4af4-9e19-b5b5e3db5fd8","Type":"ContainerStarted","Data":"87732cee7cb154d0af379dda8313e89ed188de33ccdfb15f74d76b3e2d673919"} Oct 10 13:21:51 crc kubenswrapper[4745]: I1010 13:21:51.611240 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" event={"ID":"238cafef-e181-4af4-9e19-b5b5e3db5fd8","Type":"ContainerStarted","Data":"73fe7d38dd1d6c29781594ebab249d77a42ee4c82d5bcd48870d4dd9e0ca9056"} Oct 10 13:21:51 crc kubenswrapper[4745]: I1010 13:21:51.611968 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:21:51 crc kubenswrapper[4745]: I1010 13:21:51.632012 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" podStartSLOduration=27.631990866 podStartE2EDuration="27.631990866s" podCreationTimestamp="2025-10-10 13:21:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:21:51.628760612 +0000 UTC m=+225.526417385" watchObservedRunningTime="2025-10-10 13:21:51.631990866 +0000 UTC m=+225.529647629" Oct 10 13:21:51 crc kubenswrapper[4745]: I1010 13:21:51.794941 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-57864598c6-v5hh6" Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.648042 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tr9k9"] Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.648800 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tr9k9" podUID="44dedd6d-7915-4792-81c7-d8e2dc184478" containerName="registry-server" containerID="cri-o://68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b" gracePeriod=30 Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.662578 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mjlxs"] Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.663204 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mjlxs" podUID="0dbfa2cf-9426-41e9-9f46-1cfc39e2210d" containerName="registry-server" containerID="cri-o://ad127674c0af33fae60e0d970f31c7ad5702e9034a48dadd4a6ae254b18330fe" gracePeriod=30 Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.667115 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dc27t"] Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.667345 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" podUID="7bff784a-c53a-4ff1-8507-e45b2eab9f1f" containerName="marketplace-operator" containerID="cri-o://80f2e5bc5bafdcc394b34c0d4e4040990a8041ae41690054a1ed88561d193467" gracePeriod=30 Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.670580 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tb4mp"] Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.670792 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tb4mp" podUID="c8674419-b98d-4f7e-ae45-b20aca88dcd5" containerName="registry-server" containerID="cri-o://0576a9b692231d5a65ac55573a29dd12d44a79a12bd786d50f7c6a29aeb181e1" gracePeriod=30 Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.685650 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hbrh4"] Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.686365 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hbrh4" Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.689919 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a87ec403-812a-4af1-83eb-95788143fcc4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hbrh4\" (UID: \"a87ec403-812a-4af1-83eb-95788143fcc4\") " pod="openshift-marketplace/marketplace-operator-79b997595-hbrh4" Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.689994 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a87ec403-812a-4af1-83eb-95788143fcc4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hbrh4\" (UID: \"a87ec403-812a-4af1-83eb-95788143fcc4\") " pod="openshift-marketplace/marketplace-operator-79b997595-hbrh4" Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.690035 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2p2r\" (UniqueName: \"kubernetes.io/projected/a87ec403-812a-4af1-83eb-95788143fcc4-kube-api-access-z2p2r\") pod \"marketplace-operator-79b997595-hbrh4\" (UID: \"a87ec403-812a-4af1-83eb-95788143fcc4\") " pod="openshift-marketplace/marketplace-operator-79b997595-hbrh4" Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.695633 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7vxfb"] Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.695869 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7vxfb" podUID="0dd61fe6-4685-4ce4-9c35-6465d1aa666c" containerName="registry-server" containerID="cri-o://a6f69285fd5a703023ac45999dd51549510b3fcb2188ac24fabc3110fa846ced" gracePeriod=30 Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.703888 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hbrh4"] Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.791246 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a87ec403-812a-4af1-83eb-95788143fcc4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hbrh4\" (UID: \"a87ec403-812a-4af1-83eb-95788143fcc4\") " pod="openshift-marketplace/marketplace-operator-79b997595-hbrh4" Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.791318 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2p2r\" (UniqueName: \"kubernetes.io/projected/a87ec403-812a-4af1-83eb-95788143fcc4-kube-api-access-z2p2r\") pod \"marketplace-operator-79b997595-hbrh4\" (UID: \"a87ec403-812a-4af1-83eb-95788143fcc4\") " pod="openshift-marketplace/marketplace-operator-79b997595-hbrh4" Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.791347 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a87ec403-812a-4af1-83eb-95788143fcc4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hbrh4\" (UID: \"a87ec403-812a-4af1-83eb-95788143fcc4\") " pod="openshift-marketplace/marketplace-operator-79b997595-hbrh4" Oct 10 13:22:02 crc kubenswrapper[4745]: E1010 13:22:02.792209 4745 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b is running failed: container process not found" containerID="68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b" cmd=["grpc_health_probe","-addr=:50051"] Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.792686 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a87ec403-812a-4af1-83eb-95788143fcc4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hbrh4\" (UID: \"a87ec403-812a-4af1-83eb-95788143fcc4\") " pod="openshift-marketplace/marketplace-operator-79b997595-hbrh4" Oct 10 13:22:02 crc kubenswrapper[4745]: E1010 13:22:02.792784 4745 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b is running failed: container process not found" containerID="68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b" cmd=["grpc_health_probe","-addr=:50051"] Oct 10 13:22:02 crc kubenswrapper[4745]: E1010 13:22:02.793029 4745 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b is running failed: container process not found" containerID="68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b" cmd=["grpc_health_probe","-addr=:50051"] Oct 10 13:22:02 crc kubenswrapper[4745]: E1010 13:22:02.793076 4745 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-tr9k9" podUID="44dedd6d-7915-4792-81c7-d8e2dc184478" containerName="registry-server" Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.796988 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a87ec403-812a-4af1-83eb-95788143fcc4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hbrh4\" (UID: \"a87ec403-812a-4af1-83eb-95788143fcc4\") " pod="openshift-marketplace/marketplace-operator-79b997595-hbrh4" Oct 10 13:22:02 crc kubenswrapper[4745]: I1010 13:22:02.807328 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2p2r\" (UniqueName: \"kubernetes.io/projected/a87ec403-812a-4af1-83eb-95788143fcc4-kube-api-access-z2p2r\") pod \"marketplace-operator-79b997595-hbrh4\" (UID: \"a87ec403-812a-4af1-83eb-95788143fcc4\") " pod="openshift-marketplace/marketplace-operator-79b997595-hbrh4" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.039380 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hbrh4" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.050464 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mjlxs" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.054435 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tb4mp" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.061766 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tr9k9" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.066980 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.094531 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddcwd\" (UniqueName: \"kubernetes.io/projected/c8674419-b98d-4f7e-ae45-b20aca88dcd5-kube-api-access-ddcwd\") pod \"c8674419-b98d-4f7e-ae45-b20aca88dcd5\" (UID: \"c8674419-b98d-4f7e-ae45-b20aca88dcd5\") " Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.094571 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-marketplace-operator-metrics\") pod \"7bff784a-c53a-4ff1-8507-e45b2eab9f1f\" (UID: \"7bff784a-c53a-4ff1-8507-e45b2eab9f1f\") " Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.094593 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-marketplace-trusted-ca\") pod \"7bff784a-c53a-4ff1-8507-e45b2eab9f1f\" (UID: \"7bff784a-c53a-4ff1-8507-e45b2eab9f1f\") " Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.094628 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44dedd6d-7915-4792-81c7-d8e2dc184478-catalog-content\") pod \"44dedd6d-7915-4792-81c7-d8e2dc184478\" (UID: \"44dedd6d-7915-4792-81c7-d8e2dc184478\") " Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.094658 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cklnn\" (UniqueName: \"kubernetes.io/projected/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-kube-api-access-cklnn\") pod \"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d\" (UID: \"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d\") " Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.094683 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8674419-b98d-4f7e-ae45-b20aca88dcd5-utilities\") pod \"c8674419-b98d-4f7e-ae45-b20aca88dcd5\" (UID: \"c8674419-b98d-4f7e-ae45-b20aca88dcd5\") " Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.094711 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czm9p\" (UniqueName: \"kubernetes.io/projected/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-kube-api-access-czm9p\") pod \"7bff784a-c53a-4ff1-8507-e45b2eab9f1f\" (UID: \"7bff784a-c53a-4ff1-8507-e45b2eab9f1f\") " Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.094806 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-utilities\") pod \"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d\" (UID: \"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d\") " Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.094839 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8674419-b98d-4f7e-ae45-b20aca88dcd5-catalog-content\") pod \"c8674419-b98d-4f7e-ae45-b20aca88dcd5\" (UID: \"c8674419-b98d-4f7e-ae45-b20aca88dcd5\") " Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.094874 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-catalog-content\") pod \"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d\" (UID: \"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d\") " Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.094895 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krclq\" (UniqueName: \"kubernetes.io/projected/44dedd6d-7915-4792-81c7-d8e2dc184478-kube-api-access-krclq\") pod \"44dedd6d-7915-4792-81c7-d8e2dc184478\" (UID: \"44dedd6d-7915-4792-81c7-d8e2dc184478\") " Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.094915 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44dedd6d-7915-4792-81c7-d8e2dc184478-utilities\") pod \"44dedd6d-7915-4792-81c7-d8e2dc184478\" (UID: \"44dedd6d-7915-4792-81c7-d8e2dc184478\") " Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.095967 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "7bff784a-c53a-4ff1-8507-e45b2eab9f1f" (UID: "7bff784a-c53a-4ff1-8507-e45b2eab9f1f"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.096294 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8674419-b98d-4f7e-ae45-b20aca88dcd5-utilities" (OuterVolumeSpecName: "utilities") pod "c8674419-b98d-4f7e-ae45-b20aca88dcd5" (UID: "c8674419-b98d-4f7e-ae45-b20aca88dcd5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.096506 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44dedd6d-7915-4792-81c7-d8e2dc184478-utilities" (OuterVolumeSpecName: "utilities") pod "44dedd6d-7915-4792-81c7-d8e2dc184478" (UID: "44dedd6d-7915-4792-81c7-d8e2dc184478"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.097469 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-utilities" (OuterVolumeSpecName: "utilities") pod "0dbfa2cf-9426-41e9-9f46-1cfc39e2210d" (UID: "0dbfa2cf-9426-41e9-9f46-1cfc39e2210d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.100801 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44dedd6d-7915-4792-81c7-d8e2dc184478-kube-api-access-krclq" (OuterVolumeSpecName: "kube-api-access-krclq") pod "44dedd6d-7915-4792-81c7-d8e2dc184478" (UID: "44dedd6d-7915-4792-81c7-d8e2dc184478"). InnerVolumeSpecName "kube-api-access-krclq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.101719 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "7bff784a-c53a-4ff1-8507-e45b2eab9f1f" (UID: "7bff784a-c53a-4ff1-8507-e45b2eab9f1f"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.102821 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8674419-b98d-4f7e-ae45-b20aca88dcd5-kube-api-access-ddcwd" (OuterVolumeSpecName: "kube-api-access-ddcwd") pod "c8674419-b98d-4f7e-ae45-b20aca88dcd5" (UID: "c8674419-b98d-4f7e-ae45-b20aca88dcd5"). InnerVolumeSpecName "kube-api-access-ddcwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.103967 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-kube-api-access-czm9p" (OuterVolumeSpecName: "kube-api-access-czm9p") pod "7bff784a-c53a-4ff1-8507-e45b2eab9f1f" (UID: "7bff784a-c53a-4ff1-8507-e45b2eab9f1f"). InnerVolumeSpecName "kube-api-access-czm9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.105661 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-kube-api-access-cklnn" (OuterVolumeSpecName: "kube-api-access-cklnn") pod "0dbfa2cf-9426-41e9-9f46-1cfc39e2210d" (UID: "0dbfa2cf-9426-41e9-9f46-1cfc39e2210d"). InnerVolumeSpecName "kube-api-access-cklnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.138877 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7vxfb" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.160096 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44dedd6d-7915-4792-81c7-d8e2dc184478-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44dedd6d-7915-4792-81c7-d8e2dc184478" (UID: "44dedd6d-7915-4792-81c7-d8e2dc184478"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.168127 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8674419-b98d-4f7e-ae45-b20aca88dcd5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8674419-b98d-4f7e-ae45-b20aca88dcd5" (UID: "c8674419-b98d-4f7e-ae45-b20aca88dcd5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.196095 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-catalog-content\") pod \"0dd61fe6-4685-4ce4-9c35-6465d1aa666c\" (UID: \"0dd61fe6-4685-4ce4-9c35-6465d1aa666c\") " Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.196222 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff27t\" (UniqueName: \"kubernetes.io/projected/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-kube-api-access-ff27t\") pod \"0dd61fe6-4685-4ce4-9c35-6465d1aa666c\" (UID: \"0dd61fe6-4685-4ce4-9c35-6465d1aa666c\") " Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.196298 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-utilities\") pod \"0dd61fe6-4685-4ce4-9c35-6465d1aa666c\" (UID: \"0dd61fe6-4685-4ce4-9c35-6465d1aa666c\") " Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.196495 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8674419-b98d-4f7e-ae45-b20aca88dcd5-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.196507 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czm9p\" (UniqueName: \"kubernetes.io/projected/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-kube-api-access-czm9p\") on node \"crc\" DevicePath \"\"" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.196516 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.196525 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8674419-b98d-4f7e-ae45-b20aca88dcd5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.196534 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krclq\" (UniqueName: \"kubernetes.io/projected/44dedd6d-7915-4792-81c7-d8e2dc184478-kube-api-access-krclq\") on node \"crc\" DevicePath \"\"" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.196542 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44dedd6d-7915-4792-81c7-d8e2dc184478-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.196550 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddcwd\" (UniqueName: \"kubernetes.io/projected/c8674419-b98d-4f7e-ae45-b20aca88dcd5-kube-api-access-ddcwd\") on node \"crc\" DevicePath \"\"" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.196558 4745 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.196566 4745 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7bff784a-c53a-4ff1-8507-e45b2eab9f1f-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.196575 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44dedd6d-7915-4792-81c7-d8e2dc184478-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.196583 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cklnn\" (UniqueName: \"kubernetes.io/projected/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-kube-api-access-cklnn\") on node \"crc\" DevicePath \"\"" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.197828 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-utilities" (OuterVolumeSpecName: "utilities") pod "0dd61fe6-4685-4ce4-9c35-6465d1aa666c" (UID: "0dd61fe6-4685-4ce4-9c35-6465d1aa666c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.198806 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-kube-api-access-ff27t" (OuterVolumeSpecName: "kube-api-access-ff27t") pod "0dd61fe6-4685-4ce4-9c35-6465d1aa666c" (UID: "0dd61fe6-4685-4ce4-9c35-6465d1aa666c"). InnerVolumeSpecName "kube-api-access-ff27t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.201999 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0dbfa2cf-9426-41e9-9f46-1cfc39e2210d" (UID: "0dbfa2cf-9426-41e9-9f46-1cfc39e2210d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.261841 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hbrh4"] Oct 10 13:22:03 crc kubenswrapper[4745]: W1010 13:22:03.268230 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda87ec403_812a_4af1_83eb_95788143fcc4.slice/crio-e988bdb802756dc6b8718f1c7ffcbc909538871234685f0ab8180726fbe6f0f6 WatchSource:0}: Error finding container e988bdb802756dc6b8718f1c7ffcbc909538871234685f0ab8180726fbe6f0f6: Status 404 returned error can't find the container with id e988bdb802756dc6b8718f1c7ffcbc909538871234685f0ab8180726fbe6f0f6 Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.294377 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0dd61fe6-4685-4ce4-9c35-6465d1aa666c" (UID: "0dd61fe6-4685-4ce4-9c35-6465d1aa666c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.298298 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.298342 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff27t\" (UniqueName: \"kubernetes.io/projected/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-kube-api-access-ff27t\") on node \"crc\" DevicePath \"\"" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.298358 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.298372 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd61fe6-4685-4ce4-9c35-6465d1aa666c-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.688493 4745 generic.go:334] "Generic (PLEG): container finished" podID="c8674419-b98d-4f7e-ae45-b20aca88dcd5" containerID="0576a9b692231d5a65ac55573a29dd12d44a79a12bd786d50f7c6a29aeb181e1" exitCode=0 Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.688536 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tb4mp" event={"ID":"c8674419-b98d-4f7e-ae45-b20aca88dcd5","Type":"ContainerDied","Data":"0576a9b692231d5a65ac55573a29dd12d44a79a12bd786d50f7c6a29aeb181e1"} Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.688590 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tb4mp" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.690024 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tb4mp" event={"ID":"c8674419-b98d-4f7e-ae45-b20aca88dcd5","Type":"ContainerDied","Data":"42109b9143c8be02196abe79d359bfafe4b0d9290ff470fac9333db657d89305"} Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.690188 4745 scope.go:117] "RemoveContainer" containerID="0576a9b692231d5a65ac55573a29dd12d44a79a12bd786d50f7c6a29aeb181e1" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.691939 4745 generic.go:334] "Generic (PLEG): container finished" podID="0dd61fe6-4685-4ce4-9c35-6465d1aa666c" containerID="a6f69285fd5a703023ac45999dd51549510b3fcb2188ac24fabc3110fa846ced" exitCode=0 Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.691974 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7vxfb" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.691997 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vxfb" event={"ID":"0dd61fe6-4685-4ce4-9c35-6465d1aa666c","Type":"ContainerDied","Data":"a6f69285fd5a703023ac45999dd51549510b3fcb2188ac24fabc3110fa846ced"} Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.693144 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vxfb" event={"ID":"0dd61fe6-4685-4ce4-9c35-6465d1aa666c","Type":"ContainerDied","Data":"2cb0c72220cae02f3c289ec548884e64429e7921bbb7ecc07ef4d11da7c20b3b"} Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.698461 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hbrh4" event={"ID":"a87ec403-812a-4af1-83eb-95788143fcc4","Type":"ContainerStarted","Data":"aef26b289c2536e2e43197088f95d05cf4a3305e315299371ff6de8f009d3dce"} Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.698608 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-hbrh4" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.698769 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hbrh4" event={"ID":"a87ec403-812a-4af1-83eb-95788143fcc4","Type":"ContainerStarted","Data":"e988bdb802756dc6b8718f1c7ffcbc909538871234685f0ab8180726fbe6f0f6"} Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.700614 4745 generic.go:334] "Generic (PLEG): container finished" podID="44dedd6d-7915-4792-81c7-d8e2dc184478" containerID="68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b" exitCode=0 Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.700665 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tr9k9" event={"ID":"44dedd6d-7915-4792-81c7-d8e2dc184478","Type":"ContainerDied","Data":"68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b"} Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.700683 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tr9k9" event={"ID":"44dedd6d-7915-4792-81c7-d8e2dc184478","Type":"ContainerDied","Data":"e3f76a1b7fc7670e96525f9edc699918cb36bf5e9864e3241f3f15b77ff17494"} Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.700687 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tr9k9" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.701786 4745 generic.go:334] "Generic (PLEG): container finished" podID="7bff784a-c53a-4ff1-8507-e45b2eab9f1f" containerID="80f2e5bc5bafdcc394b34c0d4e4040990a8041ae41690054a1ed88561d193467" exitCode=0 Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.701827 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" event={"ID":"7bff784a-c53a-4ff1-8507-e45b2eab9f1f","Type":"ContainerDied","Data":"80f2e5bc5bafdcc394b34c0d4e4040990a8041ae41690054a1ed88561d193467"} Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.701841 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" event={"ID":"7bff784a-c53a-4ff1-8507-e45b2eab9f1f","Type":"ContainerDied","Data":"50cfadd13bc8e509aee08d0f5bfd91be36723989cb7f05ed99a76a48338a9805"} Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.701894 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dc27t" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.702292 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-hbrh4" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.708625 4745 generic.go:334] "Generic (PLEG): container finished" podID="0dbfa2cf-9426-41e9-9f46-1cfc39e2210d" containerID="ad127674c0af33fae60e0d970f31c7ad5702e9034a48dadd4a6ae254b18330fe" exitCode=0 Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.708655 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjlxs" event={"ID":"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d","Type":"ContainerDied","Data":"ad127674c0af33fae60e0d970f31c7ad5702e9034a48dadd4a6ae254b18330fe"} Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.708673 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjlxs" event={"ID":"0dbfa2cf-9426-41e9-9f46-1cfc39e2210d","Type":"ContainerDied","Data":"c8fdf7017342861a13d882ac126123dce706c30e5872f89277bcf693a92951be"} Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.708749 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mjlxs" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.709001 4745 scope.go:117] "RemoveContainer" containerID="6ff7678a4d663ac651efab7ce6c01c82a34c0d1ece1e6cb83886015454c984e6" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.717081 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-hbrh4" podStartSLOduration=1.717058836 podStartE2EDuration="1.717058836s" podCreationTimestamp="2025-10-10 13:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:22:03.712940074 +0000 UTC m=+237.610596847" watchObservedRunningTime="2025-10-10 13:22:03.717058836 +0000 UTC m=+237.614715629" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.728416 4745 scope.go:117] "RemoveContainer" containerID="e41a4172924e16e15fcb393fb53b117ef007cb4fa979524b57b0cc12af705814" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.767040 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tb4mp"] Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.771241 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tb4mp"] Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.777557 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dc27t"] Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.778491 4745 scope.go:117] "RemoveContainer" containerID="0576a9b692231d5a65ac55573a29dd12d44a79a12bd786d50f7c6a29aeb181e1" Oct 10 13:22:03 crc kubenswrapper[4745]: E1010 13:22:03.779692 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0576a9b692231d5a65ac55573a29dd12d44a79a12bd786d50f7c6a29aeb181e1\": container with ID starting with 0576a9b692231d5a65ac55573a29dd12d44a79a12bd786d50f7c6a29aeb181e1 not found: ID does not exist" containerID="0576a9b692231d5a65ac55573a29dd12d44a79a12bd786d50f7c6a29aeb181e1" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.779740 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0576a9b692231d5a65ac55573a29dd12d44a79a12bd786d50f7c6a29aeb181e1"} err="failed to get container status \"0576a9b692231d5a65ac55573a29dd12d44a79a12bd786d50f7c6a29aeb181e1\": rpc error: code = NotFound desc = could not find container \"0576a9b692231d5a65ac55573a29dd12d44a79a12bd786d50f7c6a29aeb181e1\": container with ID starting with 0576a9b692231d5a65ac55573a29dd12d44a79a12bd786d50f7c6a29aeb181e1 not found: ID does not exist" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.779784 4745 scope.go:117] "RemoveContainer" containerID="6ff7678a4d663ac651efab7ce6c01c82a34c0d1ece1e6cb83886015454c984e6" Oct 10 13:22:03 crc kubenswrapper[4745]: E1010 13:22:03.780179 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ff7678a4d663ac651efab7ce6c01c82a34c0d1ece1e6cb83886015454c984e6\": container with ID starting with 6ff7678a4d663ac651efab7ce6c01c82a34c0d1ece1e6cb83886015454c984e6 not found: ID does not exist" containerID="6ff7678a4d663ac651efab7ce6c01c82a34c0d1ece1e6cb83886015454c984e6" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.780232 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ff7678a4d663ac651efab7ce6c01c82a34c0d1ece1e6cb83886015454c984e6"} err="failed to get container status \"6ff7678a4d663ac651efab7ce6c01c82a34c0d1ece1e6cb83886015454c984e6\": rpc error: code = NotFound desc = could not find container \"6ff7678a4d663ac651efab7ce6c01c82a34c0d1ece1e6cb83886015454c984e6\": container with ID starting with 6ff7678a4d663ac651efab7ce6c01c82a34c0d1ece1e6cb83886015454c984e6 not found: ID does not exist" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.780270 4745 scope.go:117] "RemoveContainer" containerID="e41a4172924e16e15fcb393fb53b117ef007cb4fa979524b57b0cc12af705814" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.781023 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dc27t"] Oct 10 13:22:03 crc kubenswrapper[4745]: E1010 13:22:03.781384 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e41a4172924e16e15fcb393fb53b117ef007cb4fa979524b57b0cc12af705814\": container with ID starting with e41a4172924e16e15fcb393fb53b117ef007cb4fa979524b57b0cc12af705814 not found: ID does not exist" containerID="e41a4172924e16e15fcb393fb53b117ef007cb4fa979524b57b0cc12af705814" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.781542 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e41a4172924e16e15fcb393fb53b117ef007cb4fa979524b57b0cc12af705814"} err="failed to get container status \"e41a4172924e16e15fcb393fb53b117ef007cb4fa979524b57b0cc12af705814\": rpc error: code = NotFound desc = could not find container \"e41a4172924e16e15fcb393fb53b117ef007cb4fa979524b57b0cc12af705814\": container with ID starting with e41a4172924e16e15fcb393fb53b117ef007cb4fa979524b57b0cc12af705814 not found: ID does not exist" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.781582 4745 scope.go:117] "RemoveContainer" containerID="a6f69285fd5a703023ac45999dd51549510b3fcb2188ac24fabc3110fa846ced" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.797526 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tr9k9"] Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.814150 4745 scope.go:117] "RemoveContainer" containerID="44d1405f92c2600ab57da22329b2c006d505868286158779c7f2be57ab67886a" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.815056 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tr9k9"] Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.818606 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7vxfb"] Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.821760 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7vxfb"] Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.824793 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mjlxs"] Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.827200 4745 scope.go:117] "RemoveContainer" containerID="0d61d2539a10e78356994e3119e9f5223131f138e460508848bebb59ca5314e8" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.827738 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mjlxs"] Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.838940 4745 scope.go:117] "RemoveContainer" containerID="a6f69285fd5a703023ac45999dd51549510b3fcb2188ac24fabc3110fa846ced" Oct 10 13:22:03 crc kubenswrapper[4745]: E1010 13:22:03.839299 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6f69285fd5a703023ac45999dd51549510b3fcb2188ac24fabc3110fa846ced\": container with ID starting with a6f69285fd5a703023ac45999dd51549510b3fcb2188ac24fabc3110fa846ced not found: ID does not exist" containerID="a6f69285fd5a703023ac45999dd51549510b3fcb2188ac24fabc3110fa846ced" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.839327 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6f69285fd5a703023ac45999dd51549510b3fcb2188ac24fabc3110fa846ced"} err="failed to get container status \"a6f69285fd5a703023ac45999dd51549510b3fcb2188ac24fabc3110fa846ced\": rpc error: code = NotFound desc = could not find container \"a6f69285fd5a703023ac45999dd51549510b3fcb2188ac24fabc3110fa846ced\": container with ID starting with a6f69285fd5a703023ac45999dd51549510b3fcb2188ac24fabc3110fa846ced not found: ID does not exist" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.839348 4745 scope.go:117] "RemoveContainer" containerID="44d1405f92c2600ab57da22329b2c006d505868286158779c7f2be57ab67886a" Oct 10 13:22:03 crc kubenswrapper[4745]: E1010 13:22:03.839563 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44d1405f92c2600ab57da22329b2c006d505868286158779c7f2be57ab67886a\": container with ID starting with 44d1405f92c2600ab57da22329b2c006d505868286158779c7f2be57ab67886a not found: ID does not exist" containerID="44d1405f92c2600ab57da22329b2c006d505868286158779c7f2be57ab67886a" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.839581 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44d1405f92c2600ab57da22329b2c006d505868286158779c7f2be57ab67886a"} err="failed to get container status \"44d1405f92c2600ab57da22329b2c006d505868286158779c7f2be57ab67886a\": rpc error: code = NotFound desc = could not find container \"44d1405f92c2600ab57da22329b2c006d505868286158779c7f2be57ab67886a\": container with ID starting with 44d1405f92c2600ab57da22329b2c006d505868286158779c7f2be57ab67886a not found: ID does not exist" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.839594 4745 scope.go:117] "RemoveContainer" containerID="0d61d2539a10e78356994e3119e9f5223131f138e460508848bebb59ca5314e8" Oct 10 13:22:03 crc kubenswrapper[4745]: E1010 13:22:03.839856 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d61d2539a10e78356994e3119e9f5223131f138e460508848bebb59ca5314e8\": container with ID starting with 0d61d2539a10e78356994e3119e9f5223131f138e460508848bebb59ca5314e8 not found: ID does not exist" containerID="0d61d2539a10e78356994e3119e9f5223131f138e460508848bebb59ca5314e8" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.839901 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d61d2539a10e78356994e3119e9f5223131f138e460508848bebb59ca5314e8"} err="failed to get container status \"0d61d2539a10e78356994e3119e9f5223131f138e460508848bebb59ca5314e8\": rpc error: code = NotFound desc = could not find container \"0d61d2539a10e78356994e3119e9f5223131f138e460508848bebb59ca5314e8\": container with ID starting with 0d61d2539a10e78356994e3119e9f5223131f138e460508848bebb59ca5314e8 not found: ID does not exist" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.839934 4745 scope.go:117] "RemoveContainer" containerID="68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.852486 4745 scope.go:117] "RemoveContainer" containerID="5c805cbeec106e57e402a0564d9d743affa885c716c81c33e95c2f081d07c08a" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.865455 4745 scope.go:117] "RemoveContainer" containerID="3d37cb7176fc1f6e218e9c36d9663913ea1f40a50e464386ff304905f119272d" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.877553 4745 scope.go:117] "RemoveContainer" containerID="68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b" Oct 10 13:22:03 crc kubenswrapper[4745]: E1010 13:22:03.877905 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b\": container with ID starting with 68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b not found: ID does not exist" containerID="68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.877942 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b"} err="failed to get container status \"68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b\": rpc error: code = NotFound desc = could not find container \"68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b\": container with ID starting with 68ac458bba606c15cb7aaa2f857c67fc26a73b18bd15235eda5c9243e5ac566b not found: ID does not exist" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.877967 4745 scope.go:117] "RemoveContainer" containerID="5c805cbeec106e57e402a0564d9d743affa885c716c81c33e95c2f081d07c08a" Oct 10 13:22:03 crc kubenswrapper[4745]: E1010 13:22:03.878232 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c805cbeec106e57e402a0564d9d743affa885c716c81c33e95c2f081d07c08a\": container with ID starting with 5c805cbeec106e57e402a0564d9d743affa885c716c81c33e95c2f081d07c08a not found: ID does not exist" containerID="5c805cbeec106e57e402a0564d9d743affa885c716c81c33e95c2f081d07c08a" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.878261 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c805cbeec106e57e402a0564d9d743affa885c716c81c33e95c2f081d07c08a"} err="failed to get container status \"5c805cbeec106e57e402a0564d9d743affa885c716c81c33e95c2f081d07c08a\": rpc error: code = NotFound desc = could not find container \"5c805cbeec106e57e402a0564d9d743affa885c716c81c33e95c2f081d07c08a\": container with ID starting with 5c805cbeec106e57e402a0564d9d743affa885c716c81c33e95c2f081d07c08a not found: ID does not exist" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.878284 4745 scope.go:117] "RemoveContainer" containerID="3d37cb7176fc1f6e218e9c36d9663913ea1f40a50e464386ff304905f119272d" Oct 10 13:22:03 crc kubenswrapper[4745]: E1010 13:22:03.878590 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d37cb7176fc1f6e218e9c36d9663913ea1f40a50e464386ff304905f119272d\": container with ID starting with 3d37cb7176fc1f6e218e9c36d9663913ea1f40a50e464386ff304905f119272d not found: ID does not exist" containerID="3d37cb7176fc1f6e218e9c36d9663913ea1f40a50e464386ff304905f119272d" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.878610 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d37cb7176fc1f6e218e9c36d9663913ea1f40a50e464386ff304905f119272d"} err="failed to get container status \"3d37cb7176fc1f6e218e9c36d9663913ea1f40a50e464386ff304905f119272d\": rpc error: code = NotFound desc = could not find container \"3d37cb7176fc1f6e218e9c36d9663913ea1f40a50e464386ff304905f119272d\": container with ID starting with 3d37cb7176fc1f6e218e9c36d9663913ea1f40a50e464386ff304905f119272d not found: ID does not exist" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.878622 4745 scope.go:117] "RemoveContainer" containerID="80f2e5bc5bafdcc394b34c0d4e4040990a8041ae41690054a1ed88561d193467" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.891856 4745 scope.go:117] "RemoveContainer" containerID="80f2e5bc5bafdcc394b34c0d4e4040990a8041ae41690054a1ed88561d193467" Oct 10 13:22:03 crc kubenswrapper[4745]: E1010 13:22:03.892102 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80f2e5bc5bafdcc394b34c0d4e4040990a8041ae41690054a1ed88561d193467\": container with ID starting with 80f2e5bc5bafdcc394b34c0d4e4040990a8041ae41690054a1ed88561d193467 not found: ID does not exist" containerID="80f2e5bc5bafdcc394b34c0d4e4040990a8041ae41690054a1ed88561d193467" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.892151 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80f2e5bc5bafdcc394b34c0d4e4040990a8041ae41690054a1ed88561d193467"} err="failed to get container status \"80f2e5bc5bafdcc394b34c0d4e4040990a8041ae41690054a1ed88561d193467\": rpc error: code = NotFound desc = could not find container \"80f2e5bc5bafdcc394b34c0d4e4040990a8041ae41690054a1ed88561d193467\": container with ID starting with 80f2e5bc5bafdcc394b34c0d4e4040990a8041ae41690054a1ed88561d193467 not found: ID does not exist" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.892171 4745 scope.go:117] "RemoveContainer" containerID="ad127674c0af33fae60e0d970f31c7ad5702e9034a48dadd4a6ae254b18330fe" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.903091 4745 scope.go:117] "RemoveContainer" containerID="28fbfe26343455ca9bc32d16a76aa4571e37028ac405e5a6e48e436a98f9a535" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.914993 4745 scope.go:117] "RemoveContainer" containerID="7c12c75367c2ec55e7a4fb89940a11df819f4e956a445de25d20773bc586438b" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.925776 4745 scope.go:117] "RemoveContainer" containerID="ad127674c0af33fae60e0d970f31c7ad5702e9034a48dadd4a6ae254b18330fe" Oct 10 13:22:03 crc kubenswrapper[4745]: E1010 13:22:03.929012 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad127674c0af33fae60e0d970f31c7ad5702e9034a48dadd4a6ae254b18330fe\": container with ID starting with ad127674c0af33fae60e0d970f31c7ad5702e9034a48dadd4a6ae254b18330fe not found: ID does not exist" containerID="ad127674c0af33fae60e0d970f31c7ad5702e9034a48dadd4a6ae254b18330fe" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.929050 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad127674c0af33fae60e0d970f31c7ad5702e9034a48dadd4a6ae254b18330fe"} err="failed to get container status \"ad127674c0af33fae60e0d970f31c7ad5702e9034a48dadd4a6ae254b18330fe\": rpc error: code = NotFound desc = could not find container \"ad127674c0af33fae60e0d970f31c7ad5702e9034a48dadd4a6ae254b18330fe\": container with ID starting with ad127674c0af33fae60e0d970f31c7ad5702e9034a48dadd4a6ae254b18330fe not found: ID does not exist" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.929077 4745 scope.go:117] "RemoveContainer" containerID="28fbfe26343455ca9bc32d16a76aa4571e37028ac405e5a6e48e436a98f9a535" Oct 10 13:22:03 crc kubenswrapper[4745]: E1010 13:22:03.929342 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28fbfe26343455ca9bc32d16a76aa4571e37028ac405e5a6e48e436a98f9a535\": container with ID starting with 28fbfe26343455ca9bc32d16a76aa4571e37028ac405e5a6e48e436a98f9a535 not found: ID does not exist" containerID="28fbfe26343455ca9bc32d16a76aa4571e37028ac405e5a6e48e436a98f9a535" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.929378 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28fbfe26343455ca9bc32d16a76aa4571e37028ac405e5a6e48e436a98f9a535"} err="failed to get container status \"28fbfe26343455ca9bc32d16a76aa4571e37028ac405e5a6e48e436a98f9a535\": rpc error: code = NotFound desc = could not find container \"28fbfe26343455ca9bc32d16a76aa4571e37028ac405e5a6e48e436a98f9a535\": container with ID starting with 28fbfe26343455ca9bc32d16a76aa4571e37028ac405e5a6e48e436a98f9a535 not found: ID does not exist" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.929403 4745 scope.go:117] "RemoveContainer" containerID="7c12c75367c2ec55e7a4fb89940a11df819f4e956a445de25d20773bc586438b" Oct 10 13:22:03 crc kubenswrapper[4745]: E1010 13:22:03.929665 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c12c75367c2ec55e7a4fb89940a11df819f4e956a445de25d20773bc586438b\": container with ID starting with 7c12c75367c2ec55e7a4fb89940a11df819f4e956a445de25d20773bc586438b not found: ID does not exist" containerID="7c12c75367c2ec55e7a4fb89940a11df819f4e956a445de25d20773bc586438b" Oct 10 13:22:03 crc kubenswrapper[4745]: I1010 13:22:03.929695 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c12c75367c2ec55e7a4fb89940a11df819f4e956a445de25d20773bc586438b"} err="failed to get container status \"7c12c75367c2ec55e7a4fb89940a11df819f4e956a445de25d20773bc586438b\": rpc error: code = NotFound desc = could not find container \"7c12c75367c2ec55e7a4fb89940a11df819f4e956a445de25d20773bc586438b\": container with ID starting with 7c12c75367c2ec55e7a4fb89940a11df819f4e956a445de25d20773bc586438b not found: ID does not exist" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.752191 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dbfa2cf-9426-41e9-9f46-1cfc39e2210d" path="/var/lib/kubelet/pods/0dbfa2cf-9426-41e9-9f46-1cfc39e2210d/volumes" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.752864 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dd61fe6-4685-4ce4-9c35-6465d1aa666c" path="/var/lib/kubelet/pods/0dd61fe6-4685-4ce4-9c35-6465d1aa666c/volumes" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.753516 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44dedd6d-7915-4792-81c7-d8e2dc184478" path="/var/lib/kubelet/pods/44dedd6d-7915-4792-81c7-d8e2dc184478/volumes" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.754769 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bff784a-c53a-4ff1-8507-e45b2eab9f1f" path="/var/lib/kubelet/pods/7bff784a-c53a-4ff1-8507-e45b2eab9f1f/volumes" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.755292 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8674419-b98d-4f7e-ae45-b20aca88dcd5" path="/var/lib/kubelet/pods/c8674419-b98d-4f7e-ae45-b20aca88dcd5/volumes" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.869211 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bwpkj"] Oct 10 13:22:04 crc kubenswrapper[4745]: E1010 13:22:04.869488 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8674419-b98d-4f7e-ae45-b20aca88dcd5" containerName="extract-utilities" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.869514 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8674419-b98d-4f7e-ae45-b20aca88dcd5" containerName="extract-utilities" Oct 10 13:22:04 crc kubenswrapper[4745]: E1010 13:22:04.869531 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dd61fe6-4685-4ce4-9c35-6465d1aa666c" containerName="registry-server" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.869544 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dd61fe6-4685-4ce4-9c35-6465d1aa666c" containerName="registry-server" Oct 10 13:22:04 crc kubenswrapper[4745]: E1010 13:22:04.869564 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dbfa2cf-9426-41e9-9f46-1cfc39e2210d" containerName="extract-utilities" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.869578 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dbfa2cf-9426-41e9-9f46-1cfc39e2210d" containerName="extract-utilities" Oct 10 13:22:04 crc kubenswrapper[4745]: E1010 13:22:04.869592 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44dedd6d-7915-4792-81c7-d8e2dc184478" containerName="extract-content" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.869604 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="44dedd6d-7915-4792-81c7-d8e2dc184478" containerName="extract-content" Oct 10 13:22:04 crc kubenswrapper[4745]: E1010 13:22:04.869621 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dd61fe6-4685-4ce4-9c35-6465d1aa666c" containerName="extract-utilities" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.869633 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dd61fe6-4685-4ce4-9c35-6465d1aa666c" containerName="extract-utilities" Oct 10 13:22:04 crc kubenswrapper[4745]: E1010 13:22:04.869649 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dbfa2cf-9426-41e9-9f46-1cfc39e2210d" containerName="extract-content" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.869660 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dbfa2cf-9426-41e9-9f46-1cfc39e2210d" containerName="extract-content" Oct 10 13:22:04 crc kubenswrapper[4745]: E1010 13:22:04.869680 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bff784a-c53a-4ff1-8507-e45b2eab9f1f" containerName="marketplace-operator" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.869693 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bff784a-c53a-4ff1-8507-e45b2eab9f1f" containerName="marketplace-operator" Oct 10 13:22:04 crc kubenswrapper[4745]: E1010 13:22:04.869713 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8674419-b98d-4f7e-ae45-b20aca88dcd5" containerName="extract-content" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.869725 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8674419-b98d-4f7e-ae45-b20aca88dcd5" containerName="extract-content" Oct 10 13:22:04 crc kubenswrapper[4745]: E1010 13:22:04.869956 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44dedd6d-7915-4792-81c7-d8e2dc184478" containerName="extract-utilities" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.869971 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="44dedd6d-7915-4792-81c7-d8e2dc184478" containerName="extract-utilities" Oct 10 13:22:04 crc kubenswrapper[4745]: E1010 13:22:04.869993 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44dedd6d-7915-4792-81c7-d8e2dc184478" containerName="registry-server" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.870005 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="44dedd6d-7915-4792-81c7-d8e2dc184478" containerName="registry-server" Oct 10 13:22:04 crc kubenswrapper[4745]: E1010 13:22:04.870024 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dbfa2cf-9426-41e9-9f46-1cfc39e2210d" containerName="registry-server" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.870036 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dbfa2cf-9426-41e9-9f46-1cfc39e2210d" containerName="registry-server" Oct 10 13:22:04 crc kubenswrapper[4745]: E1010 13:22:04.870049 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dd61fe6-4685-4ce4-9c35-6465d1aa666c" containerName="extract-content" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.870062 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dd61fe6-4685-4ce4-9c35-6465d1aa666c" containerName="extract-content" Oct 10 13:22:04 crc kubenswrapper[4745]: E1010 13:22:04.870079 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8674419-b98d-4f7e-ae45-b20aca88dcd5" containerName="registry-server" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.870091 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8674419-b98d-4f7e-ae45-b20aca88dcd5" containerName="registry-server" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.870243 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8674419-b98d-4f7e-ae45-b20aca88dcd5" containerName="registry-server" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.870266 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="44dedd6d-7915-4792-81c7-d8e2dc184478" containerName="registry-server" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.870286 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bff784a-c53a-4ff1-8507-e45b2eab9f1f" containerName="marketplace-operator" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.870304 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dd61fe6-4685-4ce4-9c35-6465d1aa666c" containerName="registry-server" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.870325 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dbfa2cf-9426-41e9-9f46-1cfc39e2210d" containerName="registry-server" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.871439 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bwpkj" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.875947 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.882400 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bwpkj"] Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.933677 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/511552f9-6c48-4742-9332-a5724f4d5697-catalog-content\") pod \"redhat-marketplace-bwpkj\" (UID: \"511552f9-6c48-4742-9332-a5724f4d5697\") " pod="openshift-marketplace/redhat-marketplace-bwpkj" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.933764 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/511552f9-6c48-4742-9332-a5724f4d5697-utilities\") pod \"redhat-marketplace-bwpkj\" (UID: \"511552f9-6c48-4742-9332-a5724f4d5697\") " pod="openshift-marketplace/redhat-marketplace-bwpkj" Oct 10 13:22:04 crc kubenswrapper[4745]: I1010 13:22:04.933843 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vmzj\" (UniqueName: \"kubernetes.io/projected/511552f9-6c48-4742-9332-a5724f4d5697-kube-api-access-4vmzj\") pod \"redhat-marketplace-bwpkj\" (UID: \"511552f9-6c48-4742-9332-a5724f4d5697\") " pod="openshift-marketplace/redhat-marketplace-bwpkj" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.034861 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/511552f9-6c48-4742-9332-a5724f4d5697-utilities\") pod \"redhat-marketplace-bwpkj\" (UID: \"511552f9-6c48-4742-9332-a5724f4d5697\") " pod="openshift-marketplace/redhat-marketplace-bwpkj" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.035378 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vmzj\" (UniqueName: \"kubernetes.io/projected/511552f9-6c48-4742-9332-a5724f4d5697-kube-api-access-4vmzj\") pod \"redhat-marketplace-bwpkj\" (UID: \"511552f9-6c48-4742-9332-a5724f4d5697\") " pod="openshift-marketplace/redhat-marketplace-bwpkj" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.035319 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/511552f9-6c48-4742-9332-a5724f4d5697-utilities\") pod \"redhat-marketplace-bwpkj\" (UID: \"511552f9-6c48-4742-9332-a5724f4d5697\") " pod="openshift-marketplace/redhat-marketplace-bwpkj" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.037040 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/511552f9-6c48-4742-9332-a5724f4d5697-catalog-content\") pod \"redhat-marketplace-bwpkj\" (UID: \"511552f9-6c48-4742-9332-a5724f4d5697\") " pod="openshift-marketplace/redhat-marketplace-bwpkj" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.037446 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/511552f9-6c48-4742-9332-a5724f4d5697-catalog-content\") pod \"redhat-marketplace-bwpkj\" (UID: \"511552f9-6c48-4742-9332-a5724f4d5697\") " pod="openshift-marketplace/redhat-marketplace-bwpkj" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.052061 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vmzj\" (UniqueName: \"kubernetes.io/projected/511552f9-6c48-4742-9332-a5724f4d5697-kube-api-access-4vmzj\") pod \"redhat-marketplace-bwpkj\" (UID: \"511552f9-6c48-4742-9332-a5724f4d5697\") " pod="openshift-marketplace/redhat-marketplace-bwpkj" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.068572 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-stdb8"] Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.069452 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-stdb8" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.071151 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.077278 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-stdb8"] Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.138486 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86lcb\" (UniqueName: \"kubernetes.io/projected/ba2a7571-2a26-4bf9-a642-ef927ed9c7f8-kube-api-access-86lcb\") pod \"community-operators-stdb8\" (UID: \"ba2a7571-2a26-4bf9-a642-ef927ed9c7f8\") " pod="openshift-marketplace/community-operators-stdb8" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.138532 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba2a7571-2a26-4bf9-a642-ef927ed9c7f8-utilities\") pod \"community-operators-stdb8\" (UID: \"ba2a7571-2a26-4bf9-a642-ef927ed9c7f8\") " pod="openshift-marketplace/community-operators-stdb8" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.138591 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba2a7571-2a26-4bf9-a642-ef927ed9c7f8-catalog-content\") pod \"community-operators-stdb8\" (UID: \"ba2a7571-2a26-4bf9-a642-ef927ed9c7f8\") " pod="openshift-marketplace/community-operators-stdb8" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.200013 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bwpkj" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.239805 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba2a7571-2a26-4bf9-a642-ef927ed9c7f8-catalog-content\") pod \"community-operators-stdb8\" (UID: \"ba2a7571-2a26-4bf9-a642-ef927ed9c7f8\") " pod="openshift-marketplace/community-operators-stdb8" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.239901 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86lcb\" (UniqueName: \"kubernetes.io/projected/ba2a7571-2a26-4bf9-a642-ef927ed9c7f8-kube-api-access-86lcb\") pod \"community-operators-stdb8\" (UID: \"ba2a7571-2a26-4bf9-a642-ef927ed9c7f8\") " pod="openshift-marketplace/community-operators-stdb8" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.239929 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba2a7571-2a26-4bf9-a642-ef927ed9c7f8-utilities\") pod \"community-operators-stdb8\" (UID: \"ba2a7571-2a26-4bf9-a642-ef927ed9c7f8\") " pod="openshift-marketplace/community-operators-stdb8" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.240511 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba2a7571-2a26-4bf9-a642-ef927ed9c7f8-utilities\") pod \"community-operators-stdb8\" (UID: \"ba2a7571-2a26-4bf9-a642-ef927ed9c7f8\") " pod="openshift-marketplace/community-operators-stdb8" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.240843 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba2a7571-2a26-4bf9-a642-ef927ed9c7f8-catalog-content\") pod \"community-operators-stdb8\" (UID: \"ba2a7571-2a26-4bf9-a642-ef927ed9c7f8\") " pod="openshift-marketplace/community-operators-stdb8" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.261605 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86lcb\" (UniqueName: \"kubernetes.io/projected/ba2a7571-2a26-4bf9-a642-ef927ed9c7f8-kube-api-access-86lcb\") pod \"community-operators-stdb8\" (UID: \"ba2a7571-2a26-4bf9-a642-ef927ed9c7f8\") " pod="openshift-marketplace/community-operators-stdb8" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.373493 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bwpkj"] Oct 10 13:22:05 crc kubenswrapper[4745]: W1010 13:22:05.377270 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod511552f9_6c48_4742_9332_a5724f4d5697.slice/crio-2a7fe32ccd470024438e71a891624237572dfc6e39a352d54ff81a98ebe0c3a9 WatchSource:0}: Error finding container 2a7fe32ccd470024438e71a891624237572dfc6e39a352d54ff81a98ebe0c3a9: Status 404 returned error can't find the container with id 2a7fe32ccd470024438e71a891624237572dfc6e39a352d54ff81a98ebe0c3a9 Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.402417 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-stdb8" Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.588432 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-stdb8"] Oct 10 13:22:05 crc kubenswrapper[4745]: W1010 13:22:05.599945 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba2a7571_2a26_4bf9_a642_ef927ed9c7f8.slice/crio-732ab2620942b8ddf317273d6eb3f94597ada52c6e45723c393902253948d310 WatchSource:0}: Error finding container 732ab2620942b8ddf317273d6eb3f94597ada52c6e45723c393902253948d310: Status 404 returned error can't find the container with id 732ab2620942b8ddf317273d6eb3f94597ada52c6e45723c393902253948d310 Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.726395 4745 generic.go:334] "Generic (PLEG): container finished" podID="ba2a7571-2a26-4bf9-a642-ef927ed9c7f8" containerID="77d530dcd5d11757031f7b0c478dd9cfb90e72579dc1c2ba635b0a45b5c6464b" exitCode=0 Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.726482 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stdb8" event={"ID":"ba2a7571-2a26-4bf9-a642-ef927ed9c7f8","Type":"ContainerDied","Data":"77d530dcd5d11757031f7b0c478dd9cfb90e72579dc1c2ba635b0a45b5c6464b"} Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.726659 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stdb8" event={"ID":"ba2a7571-2a26-4bf9-a642-ef927ed9c7f8","Type":"ContainerStarted","Data":"732ab2620942b8ddf317273d6eb3f94597ada52c6e45723c393902253948d310"} Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.729126 4745 generic.go:334] "Generic (PLEG): container finished" podID="511552f9-6c48-4742-9332-a5724f4d5697" containerID="67b4cf6db79b83e173efd48623730a6ed624d367273e4eb0128abfc0a571c86f" exitCode=0 Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.729400 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwpkj" event={"ID":"511552f9-6c48-4742-9332-a5724f4d5697","Type":"ContainerDied","Data":"67b4cf6db79b83e173efd48623730a6ed624d367273e4eb0128abfc0a571c86f"} Oct 10 13:22:05 crc kubenswrapper[4745]: I1010 13:22:05.729521 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwpkj" event={"ID":"511552f9-6c48-4742-9332-a5724f4d5697","Type":"ContainerStarted","Data":"2a7fe32ccd470024438e71a891624237572dfc6e39a352d54ff81a98ebe0c3a9"} Oct 10 13:22:06 crc kubenswrapper[4745]: I1010 13:22:06.740394 4745 generic.go:334] "Generic (PLEG): container finished" podID="511552f9-6c48-4742-9332-a5724f4d5697" containerID="7946c0bd4c99078ef27682d2e156119aec189f6c951a4a43dbf9d887ab1535f7" exitCode=0 Oct 10 13:22:06 crc kubenswrapper[4745]: I1010 13:22:06.740492 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwpkj" event={"ID":"511552f9-6c48-4742-9332-a5724f4d5697","Type":"ContainerDied","Data":"7946c0bd4c99078ef27682d2e156119aec189f6c951a4a43dbf9d887ab1535f7"} Oct 10 13:22:06 crc kubenswrapper[4745]: I1010 13:22:06.758791 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stdb8" event={"ID":"ba2a7571-2a26-4bf9-a642-ef927ed9c7f8","Type":"ContainerStarted","Data":"e523cfcfd52ef6dbc3d0675fc2e5d8e2e4415c100b6d42ebcbcdfe5afedb85df"} Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.278277 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h5nbz"] Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.279391 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h5nbz" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.285280 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.292524 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h5nbz"] Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.367917 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78f0f7b2-023b-4227-a64a-c3ccae73349b-catalog-content\") pod \"certified-operators-h5nbz\" (UID: \"78f0f7b2-023b-4227-a64a-c3ccae73349b\") " pod="openshift-marketplace/certified-operators-h5nbz" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.368329 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljn9d\" (UniqueName: \"kubernetes.io/projected/78f0f7b2-023b-4227-a64a-c3ccae73349b-kube-api-access-ljn9d\") pod \"certified-operators-h5nbz\" (UID: \"78f0f7b2-023b-4227-a64a-c3ccae73349b\") " pod="openshift-marketplace/certified-operators-h5nbz" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.368397 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78f0f7b2-023b-4227-a64a-c3ccae73349b-utilities\") pod \"certified-operators-h5nbz\" (UID: \"78f0f7b2-023b-4227-a64a-c3ccae73349b\") " pod="openshift-marketplace/certified-operators-h5nbz" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.469908 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljn9d\" (UniqueName: \"kubernetes.io/projected/78f0f7b2-023b-4227-a64a-c3ccae73349b-kube-api-access-ljn9d\") pod \"certified-operators-h5nbz\" (UID: \"78f0f7b2-023b-4227-a64a-c3ccae73349b\") " pod="openshift-marketplace/certified-operators-h5nbz" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.470258 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78f0f7b2-023b-4227-a64a-c3ccae73349b-utilities\") pod \"certified-operators-h5nbz\" (UID: \"78f0f7b2-023b-4227-a64a-c3ccae73349b\") " pod="openshift-marketplace/certified-operators-h5nbz" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.470318 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78f0f7b2-023b-4227-a64a-c3ccae73349b-catalog-content\") pod \"certified-operators-h5nbz\" (UID: \"78f0f7b2-023b-4227-a64a-c3ccae73349b\") " pod="openshift-marketplace/certified-operators-h5nbz" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.470706 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78f0f7b2-023b-4227-a64a-c3ccae73349b-catalog-content\") pod \"certified-operators-h5nbz\" (UID: \"78f0f7b2-023b-4227-a64a-c3ccae73349b\") " pod="openshift-marketplace/certified-operators-h5nbz" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.470014 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k7bwk"] Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.471627 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k7bwk" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.472863 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78f0f7b2-023b-4227-a64a-c3ccae73349b-utilities\") pod \"certified-operators-h5nbz\" (UID: \"78f0f7b2-023b-4227-a64a-c3ccae73349b\") " pod="openshift-marketplace/certified-operators-h5nbz" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.474020 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.486238 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k7bwk"] Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.493761 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljn9d\" (UniqueName: \"kubernetes.io/projected/78f0f7b2-023b-4227-a64a-c3ccae73349b-kube-api-access-ljn9d\") pod \"certified-operators-h5nbz\" (UID: \"78f0f7b2-023b-4227-a64a-c3ccae73349b\") " pod="openshift-marketplace/certified-operators-h5nbz" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.571369 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmg8g\" (UniqueName: \"kubernetes.io/projected/c3284f0e-0f23-4b38-bc27-3b2806155073-kube-api-access-hmg8g\") pod \"redhat-operators-k7bwk\" (UID: \"c3284f0e-0f23-4b38-bc27-3b2806155073\") " pod="openshift-marketplace/redhat-operators-k7bwk" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.571432 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3284f0e-0f23-4b38-bc27-3b2806155073-catalog-content\") pod \"redhat-operators-k7bwk\" (UID: \"c3284f0e-0f23-4b38-bc27-3b2806155073\") " pod="openshift-marketplace/redhat-operators-k7bwk" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.571487 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3284f0e-0f23-4b38-bc27-3b2806155073-utilities\") pod \"redhat-operators-k7bwk\" (UID: \"c3284f0e-0f23-4b38-bc27-3b2806155073\") " pod="openshift-marketplace/redhat-operators-k7bwk" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.611527 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h5nbz" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.675936 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmg8g\" (UniqueName: \"kubernetes.io/projected/c3284f0e-0f23-4b38-bc27-3b2806155073-kube-api-access-hmg8g\") pod \"redhat-operators-k7bwk\" (UID: \"c3284f0e-0f23-4b38-bc27-3b2806155073\") " pod="openshift-marketplace/redhat-operators-k7bwk" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.676249 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3284f0e-0f23-4b38-bc27-3b2806155073-catalog-content\") pod \"redhat-operators-k7bwk\" (UID: \"c3284f0e-0f23-4b38-bc27-3b2806155073\") " pod="openshift-marketplace/redhat-operators-k7bwk" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.676457 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3284f0e-0f23-4b38-bc27-3b2806155073-utilities\") pod \"redhat-operators-k7bwk\" (UID: \"c3284f0e-0f23-4b38-bc27-3b2806155073\") " pod="openshift-marketplace/redhat-operators-k7bwk" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.677014 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3284f0e-0f23-4b38-bc27-3b2806155073-catalog-content\") pod \"redhat-operators-k7bwk\" (UID: \"c3284f0e-0f23-4b38-bc27-3b2806155073\") " pod="openshift-marketplace/redhat-operators-k7bwk" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.677177 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3284f0e-0f23-4b38-bc27-3b2806155073-utilities\") pod \"redhat-operators-k7bwk\" (UID: \"c3284f0e-0f23-4b38-bc27-3b2806155073\") " pod="openshift-marketplace/redhat-operators-k7bwk" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.692658 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmg8g\" (UniqueName: \"kubernetes.io/projected/c3284f0e-0f23-4b38-bc27-3b2806155073-kube-api-access-hmg8g\") pod \"redhat-operators-k7bwk\" (UID: \"c3284f0e-0f23-4b38-bc27-3b2806155073\") " pod="openshift-marketplace/redhat-operators-k7bwk" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.755190 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwpkj" event={"ID":"511552f9-6c48-4742-9332-a5724f4d5697","Type":"ContainerStarted","Data":"cc0dfdd1b55cc7beca596bbbff33b7ca67845e16441d7c83e5c0b989f441838e"} Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.760907 4745 generic.go:334] "Generic (PLEG): container finished" podID="ba2a7571-2a26-4bf9-a642-ef927ed9c7f8" containerID="e523cfcfd52ef6dbc3d0675fc2e5d8e2e4415c100b6d42ebcbcdfe5afedb85df" exitCode=0 Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.760951 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stdb8" event={"ID":"ba2a7571-2a26-4bf9-a642-ef927ed9c7f8","Type":"ContainerDied","Data":"e523cfcfd52ef6dbc3d0675fc2e5d8e2e4415c100b6d42ebcbcdfe5afedb85df"} Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.772662 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bwpkj" podStartSLOduration=2.351234638 podStartE2EDuration="3.772644433s" podCreationTimestamp="2025-10-10 13:22:04 +0000 UTC" firstStartedPulling="2025-10-10 13:22:05.733065365 +0000 UTC m=+239.630722148" lastFinishedPulling="2025-10-10 13:22:07.15447518 +0000 UTC m=+241.052131943" observedRunningTime="2025-10-10 13:22:07.769516161 +0000 UTC m=+241.667172944" watchObservedRunningTime="2025-10-10 13:22:07.772644433 +0000 UTC m=+241.670301196" Oct 10 13:22:07 crc kubenswrapper[4745]: I1010 13:22:07.785068 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k7bwk" Oct 10 13:22:08 crc kubenswrapper[4745]: I1010 13:22:08.008281 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h5nbz"] Oct 10 13:22:08 crc kubenswrapper[4745]: W1010 13:22:08.020543 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78f0f7b2_023b_4227_a64a_c3ccae73349b.slice/crio-d09763cd4093eeef00cae1da223bd863e10f3e940c7791a5a634b803189ad699 WatchSource:0}: Error finding container d09763cd4093eeef00cae1da223bd863e10f3e940c7791a5a634b803189ad699: Status 404 returned error can't find the container with id d09763cd4093eeef00cae1da223bd863e10f3e940c7791a5a634b803189ad699 Oct 10 13:22:08 crc kubenswrapper[4745]: I1010 13:22:08.191659 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k7bwk"] Oct 10 13:22:08 crc kubenswrapper[4745]: I1010 13:22:08.769710 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stdb8" event={"ID":"ba2a7571-2a26-4bf9-a642-ef927ed9c7f8","Type":"ContainerStarted","Data":"733e8f9dc2e3f266dd56a62e6380a519e79632644bc636d1bc564056a8598042"} Oct 10 13:22:08 crc kubenswrapper[4745]: I1010 13:22:08.773291 4745 generic.go:334] "Generic (PLEG): container finished" podID="78f0f7b2-023b-4227-a64a-c3ccae73349b" containerID="340320a32739b1f4cdac0c0db73f25335520cfd96e5f3ef0cf9b29ed61c716ac" exitCode=0 Oct 10 13:22:08 crc kubenswrapper[4745]: I1010 13:22:08.773388 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5nbz" event={"ID":"78f0f7b2-023b-4227-a64a-c3ccae73349b","Type":"ContainerDied","Data":"340320a32739b1f4cdac0c0db73f25335520cfd96e5f3ef0cf9b29ed61c716ac"} Oct 10 13:22:08 crc kubenswrapper[4745]: I1010 13:22:08.773635 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5nbz" event={"ID":"78f0f7b2-023b-4227-a64a-c3ccae73349b","Type":"ContainerStarted","Data":"d09763cd4093eeef00cae1da223bd863e10f3e940c7791a5a634b803189ad699"} Oct 10 13:22:08 crc kubenswrapper[4745]: I1010 13:22:08.775430 4745 generic.go:334] "Generic (PLEG): container finished" podID="c3284f0e-0f23-4b38-bc27-3b2806155073" containerID="4a68050c38306f8691294d08133122b6d69ed7126d2d5e90e4f2348163565106" exitCode=0 Oct 10 13:22:08 crc kubenswrapper[4745]: I1010 13:22:08.775447 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7bwk" event={"ID":"c3284f0e-0f23-4b38-bc27-3b2806155073","Type":"ContainerDied","Data":"4a68050c38306f8691294d08133122b6d69ed7126d2d5e90e4f2348163565106"} Oct 10 13:22:08 crc kubenswrapper[4745]: I1010 13:22:08.775470 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7bwk" event={"ID":"c3284f0e-0f23-4b38-bc27-3b2806155073","Type":"ContainerStarted","Data":"6f1750ff983f1beb2014a7faa21392fcb7de5548d074029985e1a9da9a09bf6b"} Oct 10 13:22:08 crc kubenswrapper[4745]: I1010 13:22:08.787558 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-stdb8" podStartSLOduration=1.05317216 podStartE2EDuration="3.787541428s" podCreationTimestamp="2025-10-10 13:22:05 +0000 UTC" firstStartedPulling="2025-10-10 13:22:05.728344182 +0000 UTC m=+239.626000945" lastFinishedPulling="2025-10-10 13:22:08.46271345 +0000 UTC m=+242.360370213" observedRunningTime="2025-10-10 13:22:08.784896896 +0000 UTC m=+242.682553669" watchObservedRunningTime="2025-10-10 13:22:08.787541428 +0000 UTC m=+242.685198181" Oct 10 13:22:09 crc kubenswrapper[4745]: I1010 13:22:09.788772 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7bwk" event={"ID":"c3284f0e-0f23-4b38-bc27-3b2806155073","Type":"ContainerStarted","Data":"9bd445b235f201da99d18e17798321304bf4de9a5ff58466899a069542d2957f"} Oct 10 13:22:10 crc kubenswrapper[4745]: I1010 13:22:10.795624 4745 generic.go:334] "Generic (PLEG): container finished" podID="c3284f0e-0f23-4b38-bc27-3b2806155073" containerID="9bd445b235f201da99d18e17798321304bf4de9a5ff58466899a069542d2957f" exitCode=0 Oct 10 13:22:10 crc kubenswrapper[4745]: I1010 13:22:10.796435 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7bwk" event={"ID":"c3284f0e-0f23-4b38-bc27-3b2806155073","Type":"ContainerDied","Data":"9bd445b235f201da99d18e17798321304bf4de9a5ff58466899a069542d2957f"} Oct 10 13:22:11 crc kubenswrapper[4745]: I1010 13:22:11.804025 4745 generic.go:334] "Generic (PLEG): container finished" podID="78f0f7b2-023b-4227-a64a-c3ccae73349b" containerID="ad148bc37711ed7cf7aab35785e081bc233a02bb37ac2c94cbc6f29de1ae760d" exitCode=0 Oct 10 13:22:11 crc kubenswrapper[4745]: I1010 13:22:11.804070 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5nbz" event={"ID":"78f0f7b2-023b-4227-a64a-c3ccae73349b","Type":"ContainerDied","Data":"ad148bc37711ed7cf7aab35785e081bc233a02bb37ac2c94cbc6f29de1ae760d"} Oct 10 13:22:11 crc kubenswrapper[4745]: I1010 13:22:11.807884 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7bwk" event={"ID":"c3284f0e-0f23-4b38-bc27-3b2806155073","Type":"ContainerStarted","Data":"221b51c7d581dce48ad980d56e1a9e3968cca1369a962868f17ca7c110ff7837"} Oct 10 13:22:11 crc kubenswrapper[4745]: I1010 13:22:11.847543 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k7bwk" podStartSLOduration=2.238537526 podStartE2EDuration="4.84751659s" podCreationTimestamp="2025-10-10 13:22:07 +0000 UTC" firstStartedPulling="2025-10-10 13:22:08.77654962 +0000 UTC m=+242.674206383" lastFinishedPulling="2025-10-10 13:22:11.385528684 +0000 UTC m=+245.283185447" observedRunningTime="2025-10-10 13:22:11.845900788 +0000 UTC m=+245.743557551" watchObservedRunningTime="2025-10-10 13:22:11.84751659 +0000 UTC m=+245.745173393" Oct 10 13:22:12 crc kubenswrapper[4745]: I1010 13:22:12.814217 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5nbz" event={"ID":"78f0f7b2-023b-4227-a64a-c3ccae73349b","Type":"ContainerStarted","Data":"6e214f75cc2928a7c2705041787d3c13eab1959097546b8ac9d237d4431c0c90"} Oct 10 13:22:12 crc kubenswrapper[4745]: I1010 13:22:12.837090 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h5nbz" podStartSLOduration=2.3031836500000002 podStartE2EDuration="5.837072095s" podCreationTimestamp="2025-10-10 13:22:07 +0000 UTC" firstStartedPulling="2025-10-10 13:22:08.775139123 +0000 UTC m=+242.672795886" lastFinishedPulling="2025-10-10 13:22:12.309027568 +0000 UTC m=+246.206684331" observedRunningTime="2025-10-10 13:22:12.83227429 +0000 UTC m=+246.729931053" watchObservedRunningTime="2025-10-10 13:22:12.837072095 +0000 UTC m=+246.734728858" Oct 10 13:22:15 crc kubenswrapper[4745]: I1010 13:22:15.200891 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bwpkj" Oct 10 13:22:15 crc kubenswrapper[4745]: I1010 13:22:15.201899 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bwpkj" Oct 10 13:22:15 crc kubenswrapper[4745]: I1010 13:22:15.244813 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bwpkj" Oct 10 13:22:15 crc kubenswrapper[4745]: I1010 13:22:15.403721 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-stdb8" Oct 10 13:22:15 crc kubenswrapper[4745]: I1010 13:22:15.403797 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-stdb8" Oct 10 13:22:15 crc kubenswrapper[4745]: I1010 13:22:15.457168 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-stdb8" Oct 10 13:22:15 crc kubenswrapper[4745]: I1010 13:22:15.877632 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bwpkj" Oct 10 13:22:15 crc kubenswrapper[4745]: I1010 13:22:15.889482 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-stdb8" Oct 10 13:22:17 crc kubenswrapper[4745]: I1010 13:22:17.614146 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h5nbz" Oct 10 13:22:17 crc kubenswrapper[4745]: I1010 13:22:17.614201 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h5nbz" Oct 10 13:22:17 crc kubenswrapper[4745]: I1010 13:22:17.657526 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h5nbz" Oct 10 13:22:17 crc kubenswrapper[4745]: I1010 13:22:17.785651 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k7bwk" Oct 10 13:22:17 crc kubenswrapper[4745]: I1010 13:22:17.785707 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k7bwk" Oct 10 13:22:17 crc kubenswrapper[4745]: I1010 13:22:17.845467 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k7bwk" Oct 10 13:22:17 crc kubenswrapper[4745]: I1010 13:22:17.900176 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k7bwk" Oct 10 13:22:17 crc kubenswrapper[4745]: I1010 13:22:17.905340 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h5nbz" Oct 10 13:23:46 crc kubenswrapper[4745]: I1010 13:23:46.187027 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:23:46 crc kubenswrapper[4745]: I1010 13:23:46.187635 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:24:16 crc kubenswrapper[4745]: I1010 13:24:16.187343 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:24:16 crc kubenswrapper[4745]: I1010 13:24:16.188222 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.047556 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-qdqwz"] Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.048761 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.058585 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-qdqwz"] Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.210819 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcfbv\" (UniqueName: \"kubernetes.io/projected/6426acb0-8457-4772-a9f5-ab6594a6307e-kube-api-access-vcfbv\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.210876 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6426acb0-8457-4772-a9f5-ab6594a6307e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.210906 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6426acb0-8457-4772-a9f5-ab6594a6307e-registry-certificates\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.211059 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6426acb0-8457-4772-a9f5-ab6594a6307e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.211113 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6426acb0-8457-4772-a9f5-ab6594a6307e-trusted-ca\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.211227 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.211267 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6426acb0-8457-4772-a9f5-ab6594a6307e-registry-tls\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.211304 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6426acb0-8457-4772-a9f5-ab6594a6307e-bound-sa-token\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.235244 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.312932 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcfbv\" (UniqueName: \"kubernetes.io/projected/6426acb0-8457-4772-a9f5-ab6594a6307e-kube-api-access-vcfbv\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.312991 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6426acb0-8457-4772-a9f5-ab6594a6307e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.313020 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6426acb0-8457-4772-a9f5-ab6594a6307e-registry-certificates\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.313046 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6426acb0-8457-4772-a9f5-ab6594a6307e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.313070 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6426acb0-8457-4772-a9f5-ab6594a6307e-trusted-ca\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.313105 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6426acb0-8457-4772-a9f5-ab6594a6307e-registry-tls\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.313123 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6426acb0-8457-4772-a9f5-ab6594a6307e-bound-sa-token\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.315364 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6426acb0-8457-4772-a9f5-ab6594a6307e-registry-certificates\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.315571 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6426acb0-8457-4772-a9f5-ab6594a6307e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.316278 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6426acb0-8457-4772-a9f5-ab6594a6307e-trusted-ca\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.319390 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6426acb0-8457-4772-a9f5-ab6594a6307e-registry-tls\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.326475 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6426acb0-8457-4772-a9f5-ab6594a6307e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.330890 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6426acb0-8457-4772-a9f5-ab6594a6307e-bound-sa-token\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.342562 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcfbv\" (UniqueName: \"kubernetes.io/projected/6426acb0-8457-4772-a9f5-ab6594a6307e-kube-api-access-vcfbv\") pod \"image-registry-66df7c8f76-qdqwz\" (UID: \"6426acb0-8457-4772-a9f5-ab6594a6307e\") " pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.368451 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.563645 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-qdqwz"] Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.766383 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" event={"ID":"6426acb0-8457-4772-a9f5-ab6594a6307e","Type":"ContainerStarted","Data":"35a3f49b8782763bfd4c00537821c5c8f27bc2f5f5ea5618bde14e8f51cb7285"} Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.766437 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" event={"ID":"6426acb0-8457-4772-a9f5-ab6594a6307e","Type":"ContainerStarted","Data":"7bd5cdc169e836ea462c36dd3823ac2bfeb6ef40ab0c8f82b2fd1d3df8b6cc5d"} Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.766624 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:24:41 crc kubenswrapper[4745]: I1010 13:24:41.791439 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" podStartSLOduration=0.79141622 podStartE2EDuration="791.41622ms" podCreationTimestamp="2025-10-10 13:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:24:41.789091329 +0000 UTC m=+395.686748102" watchObservedRunningTime="2025-10-10 13:24:41.79141622 +0000 UTC m=+395.689072973" Oct 10 13:24:46 crc kubenswrapper[4745]: I1010 13:24:46.186978 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:24:46 crc kubenswrapper[4745]: I1010 13:24:46.188914 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:24:46 crc kubenswrapper[4745]: I1010 13:24:46.189036 4745 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:24:46 crc kubenswrapper[4745]: I1010 13:24:46.189657 4745 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"60a361e1a45479d481d0afd09183405e3656ee9875dea13f2df640bcec7617c6"} pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 13:24:46 crc kubenswrapper[4745]: I1010 13:24:46.189725 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" containerID="cri-o://60a361e1a45479d481d0afd09183405e3656ee9875dea13f2df640bcec7617c6" gracePeriod=600 Oct 10 13:24:46 crc kubenswrapper[4745]: I1010 13:24:46.803823 4745 generic.go:334] "Generic (PLEG): container finished" podID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerID="60a361e1a45479d481d0afd09183405e3656ee9875dea13f2df640bcec7617c6" exitCode=0 Oct 10 13:24:46 crc kubenswrapper[4745]: I1010 13:24:46.803971 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerDied","Data":"60a361e1a45479d481d0afd09183405e3656ee9875dea13f2df640bcec7617c6"} Oct 10 13:24:46 crc kubenswrapper[4745]: I1010 13:24:46.804275 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"c5437d8915ec61a5a5bd016aa1daf75ecd5d17ec75528801c9e54e26c6856701"} Oct 10 13:24:46 crc kubenswrapper[4745]: I1010 13:24:46.804312 4745 scope.go:117] "RemoveContainer" containerID="01add221485f9cfe2b1f584e604e3b61b3bc38303dc49e1762c04946f166cbcb" Oct 10 13:25:01 crc kubenswrapper[4745]: I1010 13:25:01.376559 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-qdqwz" Oct 10 13:25:01 crc kubenswrapper[4745]: I1010 13:25:01.462412 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qss59"] Oct 10 13:25:26 crc kubenswrapper[4745]: I1010 13:25:26.517677 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-qss59" podUID="354f8f3e-647a-4be3-a2c1-26b2cdb336b1" containerName="registry" containerID="cri-o://e667f4bfe244728ad6f95d1940e67643e4168315b728acb023d895a2a8ddf503" gracePeriod=30 Oct 10 13:25:26 crc kubenswrapper[4745]: I1010 13:25:26.880879 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.012525 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-installation-pull-secrets\") pod \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.012952 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.013030 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-trusted-ca\") pod \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.013113 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wwq8\" (UniqueName: \"kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-kube-api-access-7wwq8\") pod \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.013181 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-ca-trust-extracted\") pod \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.013228 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-registry-certificates\") pod \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.013322 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-bound-sa-token\") pod \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.013416 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-registry-tls\") pod \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\" (UID: \"354f8f3e-647a-4be3-a2c1-26b2cdb336b1\") " Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.015970 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "354f8f3e-647a-4be3-a2c1-26b2cdb336b1" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.016403 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "354f8f3e-647a-4be3-a2c1-26b2cdb336b1" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.019625 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-kube-api-access-7wwq8" (OuterVolumeSpecName: "kube-api-access-7wwq8") pod "354f8f3e-647a-4be3-a2c1-26b2cdb336b1" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1"). InnerVolumeSpecName "kube-api-access-7wwq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.020015 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "354f8f3e-647a-4be3-a2c1-26b2cdb336b1" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.021098 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "354f8f3e-647a-4be3-a2c1-26b2cdb336b1" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.021557 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "354f8f3e-647a-4be3-a2c1-26b2cdb336b1" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.025859 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "354f8f3e-647a-4be3-a2c1-26b2cdb336b1" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.040816 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "354f8f3e-647a-4be3-a2c1-26b2cdb336b1" (UID: "354f8f3e-647a-4be3-a2c1-26b2cdb336b1"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.080548 4745 generic.go:334] "Generic (PLEG): container finished" podID="354f8f3e-647a-4be3-a2c1-26b2cdb336b1" containerID="e667f4bfe244728ad6f95d1940e67643e4168315b728acb023d895a2a8ddf503" exitCode=0 Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.080612 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qss59" event={"ID":"354f8f3e-647a-4be3-a2c1-26b2cdb336b1","Type":"ContainerDied","Data":"e667f4bfe244728ad6f95d1940e67643e4168315b728acb023d895a2a8ddf503"} Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.080653 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qss59" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.080691 4745 scope.go:117] "RemoveContainer" containerID="e667f4bfe244728ad6f95d1940e67643e4168315b728acb023d895a2a8ddf503" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.080660 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qss59" event={"ID":"354f8f3e-647a-4be3-a2c1-26b2cdb336b1","Type":"ContainerDied","Data":"cd7b478275debc46ce04591111b8c4dde0276fb0f496777e745834d5df7ce6c8"} Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.099128 4745 scope.go:117] "RemoveContainer" containerID="e667f4bfe244728ad6f95d1940e67643e4168315b728acb023d895a2a8ddf503" Oct 10 13:25:27 crc kubenswrapper[4745]: E1010 13:25:27.099419 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e667f4bfe244728ad6f95d1940e67643e4168315b728acb023d895a2a8ddf503\": container with ID starting with e667f4bfe244728ad6f95d1940e67643e4168315b728acb023d895a2a8ddf503 not found: ID does not exist" containerID="e667f4bfe244728ad6f95d1940e67643e4168315b728acb023d895a2a8ddf503" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.099473 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e667f4bfe244728ad6f95d1940e67643e4168315b728acb023d895a2a8ddf503"} err="failed to get container status \"e667f4bfe244728ad6f95d1940e67643e4168315b728acb023d895a2a8ddf503\": rpc error: code = NotFound desc = could not find container \"e667f4bfe244728ad6f95d1940e67643e4168315b728acb023d895a2a8ddf503\": container with ID starting with e667f4bfe244728ad6f95d1940e67643e4168315b728acb023d895a2a8ddf503 not found: ID does not exist" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.114134 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qss59"] Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.114330 4745 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.114402 4745 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.114423 4745 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.114439 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wwq8\" (UniqueName: \"kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-kube-api-access-7wwq8\") on node \"crc\" DevicePath \"\"" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.114496 4745 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.114514 4745 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.114530 4745 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/354f8f3e-647a-4be3-a2c1-26b2cdb336b1-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 10 13:25:27 crc kubenswrapper[4745]: I1010 13:25:27.119484 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qss59"] Oct 10 13:25:28 crc kubenswrapper[4745]: I1010 13:25:28.756766 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="354f8f3e-647a-4be3-a2c1-26b2cdb336b1" path="/var/lib/kubelet/pods/354f8f3e-647a-4be3-a2c1-26b2cdb336b1/volumes" Oct 10 13:26:46 crc kubenswrapper[4745]: I1010 13:26:46.187458 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:26:46 crc kubenswrapper[4745]: I1010 13:26:46.188582 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.778997 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-rp25w"] Oct 10 13:27:12 crc kubenswrapper[4745]: E1010 13:27:12.779783 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="354f8f3e-647a-4be3-a2c1-26b2cdb336b1" containerName="registry" Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.779800 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="354f8f3e-647a-4be3-a2c1-26b2cdb336b1" containerName="registry" Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.779924 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="354f8f3e-647a-4be3-a2c1-26b2cdb336b1" containerName="registry" Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.780359 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-rp25w" Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.782692 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.782718 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.782979 4745 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-d6l6q" Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.792014 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-c24f6"] Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.792870 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-c24f6" Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.796393 4745 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-26qtp" Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.799352 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-rp25w"] Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.807373 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-c24f6"] Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.822393 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-r4mqq"] Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.823151 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-r4mqq" Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.824933 4745 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-j98sb" Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.836978 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-r4mqq"] Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.950468 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6qxd\" (UniqueName: \"kubernetes.io/projected/61fb7cf8-4f9b-4414-824f-1f5270529840-kube-api-access-n6qxd\") pod \"cert-manager-5b446d88c5-c24f6\" (UID: \"61fb7cf8-4f9b-4414-824f-1f5270529840\") " pod="cert-manager/cert-manager-5b446d88c5-c24f6" Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.950527 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w89nw\" (UniqueName: \"kubernetes.io/projected/5f2820cb-2b97-4f6f-a92d-b6f16cb83de9-kube-api-access-w89nw\") pod \"cert-manager-cainjector-7f985d654d-rp25w\" (UID: \"5f2820cb-2b97-4f6f-a92d-b6f16cb83de9\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-rp25w" Oct 10 13:27:12 crc kubenswrapper[4745]: I1010 13:27:12.950595 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jtqt\" (UniqueName: \"kubernetes.io/projected/d365e19d-2015-4d4f-935e-79a4b3998dc4-kube-api-access-5jtqt\") pod \"cert-manager-webhook-5655c58dd6-r4mqq\" (UID: \"d365e19d-2015-4d4f-935e-79a4b3998dc4\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-r4mqq" Oct 10 13:27:13 crc kubenswrapper[4745]: I1010 13:27:13.051346 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w89nw\" (UniqueName: \"kubernetes.io/projected/5f2820cb-2b97-4f6f-a92d-b6f16cb83de9-kube-api-access-w89nw\") pod \"cert-manager-cainjector-7f985d654d-rp25w\" (UID: \"5f2820cb-2b97-4f6f-a92d-b6f16cb83de9\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-rp25w" Oct 10 13:27:13 crc kubenswrapper[4745]: I1010 13:27:13.051403 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jtqt\" (UniqueName: \"kubernetes.io/projected/d365e19d-2015-4d4f-935e-79a4b3998dc4-kube-api-access-5jtqt\") pod \"cert-manager-webhook-5655c58dd6-r4mqq\" (UID: \"d365e19d-2015-4d4f-935e-79a4b3998dc4\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-r4mqq" Oct 10 13:27:13 crc kubenswrapper[4745]: I1010 13:27:13.051457 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6qxd\" (UniqueName: \"kubernetes.io/projected/61fb7cf8-4f9b-4414-824f-1f5270529840-kube-api-access-n6qxd\") pod \"cert-manager-5b446d88c5-c24f6\" (UID: \"61fb7cf8-4f9b-4414-824f-1f5270529840\") " pod="cert-manager/cert-manager-5b446d88c5-c24f6" Oct 10 13:27:13 crc kubenswrapper[4745]: I1010 13:27:13.070898 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w89nw\" (UniqueName: \"kubernetes.io/projected/5f2820cb-2b97-4f6f-a92d-b6f16cb83de9-kube-api-access-w89nw\") pod \"cert-manager-cainjector-7f985d654d-rp25w\" (UID: \"5f2820cb-2b97-4f6f-a92d-b6f16cb83de9\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-rp25w" Oct 10 13:27:13 crc kubenswrapper[4745]: I1010 13:27:13.072183 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6qxd\" (UniqueName: \"kubernetes.io/projected/61fb7cf8-4f9b-4414-824f-1f5270529840-kube-api-access-n6qxd\") pod \"cert-manager-5b446d88c5-c24f6\" (UID: \"61fb7cf8-4f9b-4414-824f-1f5270529840\") " pod="cert-manager/cert-manager-5b446d88c5-c24f6" Oct 10 13:27:13 crc kubenswrapper[4745]: I1010 13:27:13.072668 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jtqt\" (UniqueName: \"kubernetes.io/projected/d365e19d-2015-4d4f-935e-79a4b3998dc4-kube-api-access-5jtqt\") pod \"cert-manager-webhook-5655c58dd6-r4mqq\" (UID: \"d365e19d-2015-4d4f-935e-79a4b3998dc4\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-r4mqq" Oct 10 13:27:13 crc kubenswrapper[4745]: I1010 13:27:13.107525 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-rp25w" Oct 10 13:27:13 crc kubenswrapper[4745]: I1010 13:27:13.113886 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-c24f6" Oct 10 13:27:13 crc kubenswrapper[4745]: I1010 13:27:13.139344 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-r4mqq" Oct 10 13:27:13 crc kubenswrapper[4745]: I1010 13:27:13.524065 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-c24f6"] Oct 10 13:27:13 crc kubenswrapper[4745]: I1010 13:27:13.543388 4745 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 13:27:13 crc kubenswrapper[4745]: I1010 13:27:13.580850 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-rp25w"] Oct 10 13:27:13 crc kubenswrapper[4745]: W1010 13:27:13.587786 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f2820cb_2b97_4f6f_a92d_b6f16cb83de9.slice/crio-46588df623bb673e4ce1731a256ac1a7e8d2110beadb748521e1924d95d151b8 WatchSource:0}: Error finding container 46588df623bb673e4ce1731a256ac1a7e8d2110beadb748521e1924d95d151b8: Status 404 returned error can't find the container with id 46588df623bb673e4ce1731a256ac1a7e8d2110beadb748521e1924d95d151b8 Oct 10 13:27:13 crc kubenswrapper[4745]: I1010 13:27:13.593630 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-r4mqq"] Oct 10 13:27:13 crc kubenswrapper[4745]: W1010 13:27:13.600372 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd365e19d_2015_4d4f_935e_79a4b3998dc4.slice/crio-4bd37a7c0c7273130dccc9b5f8352c84e5379787960a13be287459c25844b77e WatchSource:0}: Error finding container 4bd37a7c0c7273130dccc9b5f8352c84e5379787960a13be287459c25844b77e: Status 404 returned error can't find the container with id 4bd37a7c0c7273130dccc9b5f8352c84e5379787960a13be287459c25844b77e Oct 10 13:27:13 crc kubenswrapper[4745]: I1010 13:27:13.747618 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-c24f6" event={"ID":"61fb7cf8-4f9b-4414-824f-1f5270529840","Type":"ContainerStarted","Data":"d0425d9e0781d9abd8f8cffdd4f4b047fddec451e96b84e8cb8742422b9d8aca"} Oct 10 13:27:13 crc kubenswrapper[4745]: I1010 13:27:13.748849 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-r4mqq" event={"ID":"d365e19d-2015-4d4f-935e-79a4b3998dc4","Type":"ContainerStarted","Data":"4bd37a7c0c7273130dccc9b5f8352c84e5379787960a13be287459c25844b77e"} Oct 10 13:27:13 crc kubenswrapper[4745]: I1010 13:27:13.749657 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-rp25w" event={"ID":"5f2820cb-2b97-4f6f-a92d-b6f16cb83de9","Type":"ContainerStarted","Data":"46588df623bb673e4ce1731a256ac1a7e8d2110beadb748521e1924d95d151b8"} Oct 10 13:27:16 crc kubenswrapper[4745]: I1010 13:27:16.187393 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:27:16 crc kubenswrapper[4745]: I1010 13:27:16.188247 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:27:16 crc kubenswrapper[4745]: I1010 13:27:16.770258 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-rp25w" event={"ID":"5f2820cb-2b97-4f6f-a92d-b6f16cb83de9","Type":"ContainerStarted","Data":"f7dc9390550750cb6a84b499b74e79c108cdaa944c09259fc83fb1f5b6716951"} Oct 10 13:27:16 crc kubenswrapper[4745]: I1010 13:27:16.772059 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-c24f6" event={"ID":"61fb7cf8-4f9b-4414-824f-1f5270529840","Type":"ContainerStarted","Data":"ffe74bd571b882c3f4a87ed344f59cfada83698a2eaeefcaafd3c291af593495"} Oct 10 13:27:16 crc kubenswrapper[4745]: I1010 13:27:16.823314 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-rp25w" podStartSLOduration=2.358078444 podStartE2EDuration="4.823299447s" podCreationTimestamp="2025-10-10 13:27:12 +0000 UTC" firstStartedPulling="2025-10-10 13:27:13.590795031 +0000 UTC m=+547.488451794" lastFinishedPulling="2025-10-10 13:27:16.056016034 +0000 UTC m=+549.953672797" observedRunningTime="2025-10-10 13:27:16.807447303 +0000 UTC m=+550.705104066" watchObservedRunningTime="2025-10-10 13:27:16.823299447 +0000 UTC m=+550.720956210" Oct 10 13:27:17 crc kubenswrapper[4745]: I1010 13:27:17.781300 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-r4mqq" event={"ID":"d365e19d-2015-4d4f-935e-79a4b3998dc4","Type":"ContainerStarted","Data":"7c16ca8352eba0308624e7bb492200cf00ce5e97d4237d4c4f65e44958524784"} Oct 10 13:27:17 crc kubenswrapper[4745]: I1010 13:27:17.781700 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-r4mqq" Oct 10 13:27:17 crc kubenswrapper[4745]: I1010 13:27:17.805092 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-r4mqq" podStartSLOduration=2.422133991 podStartE2EDuration="5.80506599s" podCreationTimestamp="2025-10-10 13:27:12 +0000 UTC" firstStartedPulling="2025-10-10 13:27:13.603011441 +0000 UTC m=+547.500668204" lastFinishedPulling="2025-10-10 13:27:16.98594344 +0000 UTC m=+550.883600203" observedRunningTime="2025-10-10 13:27:17.803448097 +0000 UTC m=+551.701104890" watchObservedRunningTime="2025-10-10 13:27:17.80506599 +0000 UTC m=+551.702722773" Oct 10 13:27:17 crc kubenswrapper[4745]: I1010 13:27:17.807661 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-c24f6" podStartSLOduration=3.281723028 podStartE2EDuration="5.807648377s" podCreationTimestamp="2025-10-10 13:27:12 +0000 UTC" firstStartedPulling="2025-10-10 13:27:13.543123725 +0000 UTC m=+547.440780488" lastFinishedPulling="2025-10-10 13:27:16.069049074 +0000 UTC m=+549.966705837" observedRunningTime="2025-10-10 13:27:16.824834747 +0000 UTC m=+550.722491500" watchObservedRunningTime="2025-10-10 13:27:17.807648377 +0000 UTC m=+551.705305150" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.141575 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-r4mqq" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.505928 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5bh9m"] Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.508227 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovn-controller" containerID="cri-o://70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f" gracePeriod=30 Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.508448 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="sbdb" containerID="cri-o://ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce" gracePeriod=30 Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.508360 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1" gracePeriod=30 Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.508392 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="northd" containerID="cri-o://6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6" gracePeriod=30 Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.508403 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovn-acl-logging" containerID="cri-o://982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9" gracePeriod=30 Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.508415 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="kube-rbac-proxy-node" containerID="cri-o://f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69" gracePeriod=30 Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.508254 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="nbdb" containerID="cri-o://79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210" gracePeriod=30 Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.558548 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovnkube-controller" containerID="cri-o://668fb16e5543f24cd2b93f606627d511560b74d218f48fabe602ac3781ea471d" gracePeriod=30 Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.829337 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-27d2n_1d62a30b-50a8-46a2-82c6-5d7e2f24e4de/kube-multus/1.log" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.830136 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-27d2n_1d62a30b-50a8-46a2-82c6-5d7e2f24e4de/kube-multus/0.log" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.830182 4745 generic.go:334] "Generic (PLEG): container finished" podID="1d62a30b-50a8-46a2-82c6-5d7e2f24e4de" containerID="2f78611b429d8411902427cc4f7dda27fb9f1ea97db3b312737d4bbc38d350a8" exitCode=2 Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.830232 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-27d2n" event={"ID":"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de","Type":"ContainerDied","Data":"2f78611b429d8411902427cc4f7dda27fb9f1ea97db3b312737d4bbc38d350a8"} Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.830265 4745 scope.go:117] "RemoveContainer" containerID="9f9f2cd0f5476e76620977e976038b1837ebf3b0960e3d32a4d10c9856420fdb" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.830953 4745 scope.go:117] "RemoveContainer" containerID="2f78611b429d8411902427cc4f7dda27fb9f1ea97db3b312737d4bbc38d350a8" Oct 10 13:27:23 crc kubenswrapper[4745]: E1010 13:27:23.831121 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-27d2n_openshift-multus(1d62a30b-50a8-46a2-82c6-5d7e2f24e4de)\"" pod="openshift-multus/multus-27d2n" podUID="1d62a30b-50a8-46a2-82c6-5d7e2f24e4de" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.834652 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovnkube-controller/3.log" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.837464 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovn-acl-logging/0.log" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.838211 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovn-controller/0.log" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.838918 4745 generic.go:334] "Generic (PLEG): container finished" podID="687bad4e-baff-4e09-bb33-def12c499092" containerID="668fb16e5543f24cd2b93f606627d511560b74d218f48fabe602ac3781ea471d" exitCode=0 Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.838993 4745 generic.go:334] "Generic (PLEG): container finished" podID="687bad4e-baff-4e09-bb33-def12c499092" containerID="ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce" exitCode=0 Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.839011 4745 generic.go:334] "Generic (PLEG): container finished" podID="687bad4e-baff-4e09-bb33-def12c499092" containerID="79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210" exitCode=0 Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.839068 4745 generic.go:334] "Generic (PLEG): container finished" podID="687bad4e-baff-4e09-bb33-def12c499092" containerID="6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6" exitCode=0 Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.839083 4745 generic.go:334] "Generic (PLEG): container finished" podID="687bad4e-baff-4e09-bb33-def12c499092" containerID="5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1" exitCode=0 Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.839097 4745 generic.go:334] "Generic (PLEG): container finished" podID="687bad4e-baff-4e09-bb33-def12c499092" containerID="f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69" exitCode=0 Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.839110 4745 generic.go:334] "Generic (PLEG): container finished" podID="687bad4e-baff-4e09-bb33-def12c499092" containerID="982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9" exitCode=143 Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.839162 4745 generic.go:334] "Generic (PLEG): container finished" podID="687bad4e-baff-4e09-bb33-def12c499092" containerID="70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f" exitCode=143 Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.839540 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerDied","Data":"668fb16e5543f24cd2b93f606627d511560b74d218f48fabe602ac3781ea471d"} Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.839758 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerDied","Data":"ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce"} Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.839888 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerDied","Data":"79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210"} Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.840005 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerDied","Data":"6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6"} Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.840121 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerDied","Data":"5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1"} Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.840240 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerDied","Data":"f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69"} Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.840432 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerDied","Data":"982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9"} Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.840564 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerDied","Data":"70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f"} Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.840682 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" event={"ID":"687bad4e-baff-4e09-bb33-def12c499092","Type":"ContainerDied","Data":"23496822e4dfaba11f61f672b9fa2d403b2b98b040a78aaec8f472578b98952b"} Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.840834 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23496822e4dfaba11f61f672b9fa2d403b2b98b040a78aaec8f472578b98952b" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.856014 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovnkube-controller/3.log" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.858681 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovn-acl-logging/0.log" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.859134 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovn-controller/0.log" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.859460 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.883096 4745 scope.go:117] "RemoveContainer" containerID="b6aaf80ed98f0bc57af979fdd09197f78ed844c4355943910b9de0935fbeaef8" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.929575 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nmttj"] Oct 10 13:27:23 crc kubenswrapper[4745]: E1010 13:27:23.929806 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovn-acl-logging" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.929819 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovn-acl-logging" Oct 10 13:27:23 crc kubenswrapper[4745]: E1010 13:27:23.929831 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovnkube-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.929838 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovnkube-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: E1010 13:27:23.929847 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovnkube-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.929856 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovnkube-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: E1010 13:27:23.929866 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovn-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.929875 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovn-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: E1010 13:27:23.929885 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovnkube-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.929892 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovnkube-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: E1010 13:27:23.929902 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="kubecfg-setup" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.929908 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="kubecfg-setup" Oct 10 13:27:23 crc kubenswrapper[4745]: E1010 13:27:23.929919 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="kube-rbac-proxy-node" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.929926 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="kube-rbac-proxy-node" Oct 10 13:27:23 crc kubenswrapper[4745]: E1010 13:27:23.929936 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovnkube-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.929942 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovnkube-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: E1010 13:27:23.929951 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="northd" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.929957 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="northd" Oct 10 13:27:23 crc kubenswrapper[4745]: E1010 13:27:23.929965 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="nbdb" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.929973 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="nbdb" Oct 10 13:27:23 crc kubenswrapper[4745]: E1010 13:27:23.929982 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="kube-rbac-proxy-ovn-metrics" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.929990 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="kube-rbac-proxy-ovn-metrics" Oct 10 13:27:23 crc kubenswrapper[4745]: E1010 13:27:23.930002 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="sbdb" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.930009 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="sbdb" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.930115 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovn-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.930125 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovn-acl-logging" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.930134 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="sbdb" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.930146 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="northd" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.930157 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="kube-rbac-proxy-node" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.930167 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="nbdb" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.930175 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="kube-rbac-proxy-ovn-metrics" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.930184 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovnkube-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.930195 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovnkube-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.930203 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovnkube-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.930211 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovnkube-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: E1010 13:27:23.930579 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovnkube-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.930590 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovnkube-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.930667 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="687bad4e-baff-4e09-bb33-def12c499092" containerName="ovnkube-controller" Oct 10 13:27:23 crc kubenswrapper[4745]: I1010 13:27:23.932494 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004295 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-slash\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004366 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-systemd\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004398 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-cni-netd\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004420 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-env-overrides\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004434 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-etc-openvswitch\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004456 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-ovnkube-config\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004481 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-kubelet\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004493 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-systemd-units\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004508 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-var-lib-openvswitch\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004525 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-run-ovn-kubernetes\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004541 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-ovnkube-script-lib\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004558 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-cni-bin\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004577 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9sbh\" (UniqueName: \"kubernetes.io/projected/687bad4e-baff-4e09-bb33-def12c499092-kube-api-access-l9sbh\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004596 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/687bad4e-baff-4e09-bb33-def12c499092-ovn-node-metrics-cert\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004609 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-log-socket\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004622 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-run-netns\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004643 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-var-lib-cni-networks-ovn-kubernetes\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004664 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-node-log\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004691 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-ovn\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.004705 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-openvswitch\") pod \"687bad4e-baff-4e09-bb33-def12c499092\" (UID: \"687bad4e-baff-4e09-bb33-def12c499092\") " Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.005224 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-slash" (OuterVolumeSpecName: "host-slash") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.005841 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.005837 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.005880 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.005894 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.005903 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.005886 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-log-socket" (OuterVolumeSpecName: "log-socket") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.005943 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.005942 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.005942 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-node-log" (OuterVolumeSpecName: "node-log") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.005973 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.005990 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.006024 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.006091 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.006215 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.006453 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.006480 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.010413 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/687bad4e-baff-4e09-bb33-def12c499092-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.010725 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/687bad4e-baff-4e09-bb33-def12c499092-kube-api-access-l9sbh" (OuterVolumeSpecName: "kube-api-access-l9sbh") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "kube-api-access-l9sbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.021520 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "687bad4e-baff-4e09-bb33-def12c499092" (UID: "687bad4e-baff-4e09-bb33-def12c499092"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.105775 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-run-systemd\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.105824 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-slash\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.105850 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-cni-bin\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.105884 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4d22c31e-52ef-4632-b392-91c7aa8bb03a-ovnkube-script-lib\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.105908 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d22c31e-52ef-4632-b392-91c7aa8bb03a-ovnkube-config\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.105929 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-systemd-units\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106475 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-run-openvswitch\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106557 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-cni-netd\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106580 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-etc-openvswitch\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106606 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-node-log\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106626 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-run-ovn-kubernetes\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106648 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-kubelet\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106665 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-run-netns\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106687 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-run-ovn\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106711 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fjd6\" (UniqueName: \"kubernetes.io/projected/4d22c31e-52ef-4632-b392-91c7aa8bb03a-kube-api-access-2fjd6\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106750 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d22c31e-52ef-4632-b392-91c7aa8bb03a-env-overrides\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106773 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-var-lib-openvswitch\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106800 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106824 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d22c31e-52ef-4632-b392-91c7aa8bb03a-ovn-node-metrics-cert\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106850 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-log-socket\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106896 4745 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106910 4745 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106922 4745 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106934 4745 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106945 4745 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106958 4745 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106970 4745 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106982 4745 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.106997 4745 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/687bad4e-baff-4e09-bb33-def12c499092-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.107009 4745 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.107023 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9sbh\" (UniqueName: \"kubernetes.io/projected/687bad4e-baff-4e09-bb33-def12c499092-kube-api-access-l9sbh\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.107085 4745 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/687bad4e-baff-4e09-bb33-def12c499092-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.107099 4745 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-log-socket\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.107109 4745 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.107121 4745 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.107132 4745 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-node-log\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.107145 4745 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.107156 4745 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.107168 4745 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-host-slash\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.107179 4745 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/687bad4e-baff-4e09-bb33-def12c499092-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208246 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4d22c31e-52ef-4632-b392-91c7aa8bb03a-ovnkube-script-lib\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208296 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d22c31e-52ef-4632-b392-91c7aa8bb03a-ovnkube-config\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208319 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-systemd-units\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208362 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-run-openvswitch\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208387 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-cni-netd\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208412 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-etc-openvswitch\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208434 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-node-log\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208457 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-run-ovn-kubernetes\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208479 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-kubelet\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208498 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-run-netns\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208494 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-run-openvswitch\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208521 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-run-ovn\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208557 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-run-ovn\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208578 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fjd6\" (UniqueName: \"kubernetes.io/projected/4d22c31e-52ef-4632-b392-91c7aa8bb03a-kube-api-access-2fjd6\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208600 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-cni-netd\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208625 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d22c31e-52ef-4632-b392-91c7aa8bb03a-env-overrides\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208645 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-etc-openvswitch\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208660 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-var-lib-openvswitch\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208684 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-node-log\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208705 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208720 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-run-ovn-kubernetes\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208773 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-kubelet\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208798 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d22c31e-52ef-4632-b392-91c7aa8bb03a-ovn-node-metrics-cert\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208845 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-log-socket\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208886 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-run-systemd\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208921 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-slash\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208943 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-systemd-units\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208940 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208959 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-cni-bin\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.209044 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-log-socket\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.209007 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-cni-bin\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.209048 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-run-systemd\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.209087 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-slash\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.208801 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-host-run-netns\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.209001 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d22c31e-52ef-4632-b392-91c7aa8bb03a-var-lib-openvswitch\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.209204 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4d22c31e-52ef-4632-b392-91c7aa8bb03a-ovnkube-script-lib\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.209256 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d22c31e-52ef-4632-b392-91c7aa8bb03a-ovnkube-config\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.209341 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d22c31e-52ef-4632-b392-91c7aa8bb03a-env-overrides\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.212923 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d22c31e-52ef-4632-b392-91c7aa8bb03a-ovn-node-metrics-cert\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.234640 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fjd6\" (UniqueName: \"kubernetes.io/projected/4d22c31e-52ef-4632-b392-91c7aa8bb03a-kube-api-access-2fjd6\") pod \"ovnkube-node-nmttj\" (UID: \"4d22c31e-52ef-4632-b392-91c7aa8bb03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.246241 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.849571 4745 generic.go:334] "Generic (PLEG): container finished" podID="4d22c31e-52ef-4632-b392-91c7aa8bb03a" containerID="a92fe5d84cb84b26f2ca7f1d2d32304b22960d938d964ec185d0163d51136d69" exitCode=0 Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.849662 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" event={"ID":"4d22c31e-52ef-4632-b392-91c7aa8bb03a","Type":"ContainerDied","Data":"a92fe5d84cb84b26f2ca7f1d2d32304b22960d938d964ec185d0163d51136d69"} Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.849960 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" event={"ID":"4d22c31e-52ef-4632-b392-91c7aa8bb03a","Type":"ContainerStarted","Data":"a9270c139329c8388999d7e113f3d91105b75c633546cb40ed603dae61382850"} Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.854456 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-27d2n_1d62a30b-50a8-46a2-82c6-5d7e2f24e4de/kube-multus/1.log" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.862150 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovn-acl-logging/0.log" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.863676 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bh9m_687bad4e-baff-4e09-bb33-def12c499092/ovn-controller/0.log" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.864458 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5bh9m" Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.984865 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5bh9m"] Oct 10 13:27:24 crc kubenswrapper[4745]: I1010 13:27:24.989456 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5bh9m"] Oct 10 13:27:25 crc kubenswrapper[4745]: I1010 13:27:25.873854 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" event={"ID":"4d22c31e-52ef-4632-b392-91c7aa8bb03a","Type":"ContainerStarted","Data":"913dfcbe0f456c4a4b8f8c4bba234d436569d7ebdc833d62e25624cbb74582a6"} Oct 10 13:27:25 crc kubenswrapper[4745]: I1010 13:27:25.874158 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" event={"ID":"4d22c31e-52ef-4632-b392-91c7aa8bb03a","Type":"ContainerStarted","Data":"f8293a4905c31119f834608e25a7cad17f1961ca42162f797f06d6b3339795e2"} Oct 10 13:27:25 crc kubenswrapper[4745]: I1010 13:27:25.874178 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" event={"ID":"4d22c31e-52ef-4632-b392-91c7aa8bb03a","Type":"ContainerStarted","Data":"0b88153d43ee28b02fd7c030b743c77759b7c3b2743b1dd4beb5ce8cad669c2a"} Oct 10 13:27:25 crc kubenswrapper[4745]: I1010 13:27:25.874189 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" event={"ID":"4d22c31e-52ef-4632-b392-91c7aa8bb03a","Type":"ContainerStarted","Data":"2c18a20a0705b3c0eff84fdb683bf59665a205f261d93fe857a95739471c88ad"} Oct 10 13:27:25 crc kubenswrapper[4745]: I1010 13:27:25.874198 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" event={"ID":"4d22c31e-52ef-4632-b392-91c7aa8bb03a","Type":"ContainerStarted","Data":"6a14c8b1227ed4db405338ab5438f059369306db727718895b5561b7d3534eea"} Oct 10 13:27:25 crc kubenswrapper[4745]: I1010 13:27:25.874208 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" event={"ID":"4d22c31e-52ef-4632-b392-91c7aa8bb03a","Type":"ContainerStarted","Data":"e3b2680bbe003598a2696b37f4677376ad4ed8cdb4a4f9856b53f0a194bd319e"} Oct 10 13:27:26 crc kubenswrapper[4745]: I1010 13:27:26.756103 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="687bad4e-baff-4e09-bb33-def12c499092" path="/var/lib/kubelet/pods/687bad4e-baff-4e09-bb33-def12c499092/volumes" Oct 10 13:27:28 crc kubenswrapper[4745]: I1010 13:27:28.895645 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" event={"ID":"4d22c31e-52ef-4632-b392-91c7aa8bb03a","Type":"ContainerStarted","Data":"355a3d8141d24f32dedafd200202b13306c6ad18f77206db96ece6a730908672"} Oct 10 13:27:30 crc kubenswrapper[4745]: I1010 13:27:30.913274 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" event={"ID":"4d22c31e-52ef-4632-b392-91c7aa8bb03a","Type":"ContainerStarted","Data":"80d4eefcba72aedf42d97a847c421ebc5410b61d0ee6aae5bec7861ec39032fc"} Oct 10 13:27:30 crc kubenswrapper[4745]: I1010 13:27:30.913545 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:30 crc kubenswrapper[4745]: I1010 13:27:30.913570 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:30 crc kubenswrapper[4745]: I1010 13:27:30.913714 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:30 crc kubenswrapper[4745]: I1010 13:27:30.949304 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" podStartSLOduration=7.949280386 podStartE2EDuration="7.949280386s" podCreationTimestamp="2025-10-10 13:27:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:27:30.947655723 +0000 UTC m=+564.845312526" watchObservedRunningTime="2025-10-10 13:27:30.949280386 +0000 UTC m=+564.846937159" Oct 10 13:27:30 crc kubenswrapper[4745]: I1010 13:27:30.956438 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:30 crc kubenswrapper[4745]: I1010 13:27:30.958745 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:27:36 crc kubenswrapper[4745]: I1010 13:27:36.750148 4745 scope.go:117] "RemoveContainer" containerID="2f78611b429d8411902427cc4f7dda27fb9f1ea97db3b312737d4bbc38d350a8" Oct 10 13:27:36 crc kubenswrapper[4745]: I1010 13:27:36.955946 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-27d2n_1d62a30b-50a8-46a2-82c6-5d7e2f24e4de/kube-multus/1.log" Oct 10 13:27:37 crc kubenswrapper[4745]: I1010 13:27:37.966180 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-27d2n_1d62a30b-50a8-46a2-82c6-5d7e2f24e4de/kube-multus/1.log" Oct 10 13:27:37 crc kubenswrapper[4745]: I1010 13:27:37.966268 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-27d2n" event={"ID":"1d62a30b-50a8-46a2-82c6-5d7e2f24e4de","Type":"ContainerStarted","Data":"d6b25e169b01cdbdca239d710cb22e16a840e54120086a19626fd3f9f75b2a0e"} Oct 10 13:27:46 crc kubenswrapper[4745]: I1010 13:27:46.187395 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:27:46 crc kubenswrapper[4745]: I1010 13:27:46.188425 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:27:46 crc kubenswrapper[4745]: I1010 13:27:46.188506 4745 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:27:46 crc kubenswrapper[4745]: I1010 13:27:46.190766 4745 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c5437d8915ec61a5a5bd016aa1daf75ecd5d17ec75528801c9e54e26c6856701"} pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 13:27:46 crc kubenswrapper[4745]: I1010 13:27:46.190963 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" containerID="cri-o://c5437d8915ec61a5a5bd016aa1daf75ecd5d17ec75528801c9e54e26c6856701" gracePeriod=600 Oct 10 13:27:47 crc kubenswrapper[4745]: I1010 13:27:47.030675 4745 generic.go:334] "Generic (PLEG): container finished" podID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerID="c5437d8915ec61a5a5bd016aa1daf75ecd5d17ec75528801c9e54e26c6856701" exitCode=0 Oct 10 13:27:47 crc kubenswrapper[4745]: I1010 13:27:47.031171 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerDied","Data":"c5437d8915ec61a5a5bd016aa1daf75ecd5d17ec75528801c9e54e26c6856701"} Oct 10 13:27:47 crc kubenswrapper[4745]: I1010 13:27:47.031210 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"2ba538b1feacb1e3fe2ff42d8dddb15b3e992a880af6f43548cdf890834090dc"} Oct 10 13:27:47 crc kubenswrapper[4745]: I1010 13:27:47.031233 4745 scope.go:117] "RemoveContainer" containerID="60a361e1a45479d481d0afd09183405e3656ee9875dea13f2df640bcec7617c6" Oct 10 13:27:54 crc kubenswrapper[4745]: I1010 13:27:54.281949 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nmttj" Oct 10 13:28:03 crc kubenswrapper[4745]: I1010 13:28:03.562528 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c"] Oct 10 13:28:03 crc kubenswrapper[4745]: I1010 13:28:03.564225 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" Oct 10 13:28:03 crc kubenswrapper[4745]: I1010 13:28:03.566247 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 10 13:28:03 crc kubenswrapper[4745]: I1010 13:28:03.570799 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c"] Oct 10 13:28:03 crc kubenswrapper[4745]: I1010 13:28:03.744100 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98nw7\" (UniqueName: \"kubernetes.io/projected/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-kube-api-access-98nw7\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c\" (UID: \"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" Oct 10 13:28:03 crc kubenswrapper[4745]: I1010 13:28:03.744175 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c\" (UID: \"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" Oct 10 13:28:03 crc kubenswrapper[4745]: I1010 13:28:03.744352 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c\" (UID: \"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" Oct 10 13:28:03 crc kubenswrapper[4745]: I1010 13:28:03.846190 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c\" (UID: \"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" Oct 10 13:28:03 crc kubenswrapper[4745]: I1010 13:28:03.846352 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98nw7\" (UniqueName: \"kubernetes.io/projected/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-kube-api-access-98nw7\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c\" (UID: \"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" Oct 10 13:28:03 crc kubenswrapper[4745]: I1010 13:28:03.846437 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c\" (UID: \"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" Oct 10 13:28:03 crc kubenswrapper[4745]: I1010 13:28:03.846641 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c\" (UID: \"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" Oct 10 13:28:03 crc kubenswrapper[4745]: I1010 13:28:03.847030 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c\" (UID: \"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" Oct 10 13:28:03 crc kubenswrapper[4745]: I1010 13:28:03.868476 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98nw7\" (UniqueName: \"kubernetes.io/projected/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-kube-api-access-98nw7\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c\" (UID: \"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" Oct 10 13:28:03 crc kubenswrapper[4745]: I1010 13:28:03.883275 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" Oct 10 13:28:04 crc kubenswrapper[4745]: I1010 13:28:04.323029 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c"] Oct 10 13:28:04 crc kubenswrapper[4745]: W1010 13:28:04.336054 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4cc0cf48_9860_42b1_ad6c_cbb6559ad8d4.slice/crio-98b6207c33caa39fe0cb6405cc14f2e6edcf700cbc87d56cc011fe547827f20c WatchSource:0}: Error finding container 98b6207c33caa39fe0cb6405cc14f2e6edcf700cbc87d56cc011fe547827f20c: Status 404 returned error can't find the container with id 98b6207c33caa39fe0cb6405cc14f2e6edcf700cbc87d56cc011fe547827f20c Oct 10 13:28:05 crc kubenswrapper[4745]: I1010 13:28:05.148077 4745 generic.go:334] "Generic (PLEG): container finished" podID="4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4" containerID="8d9b26dd56eae4ae0f313726c1563a99fbf04fa2706a4e0d1875576f83a54ebd" exitCode=0 Oct 10 13:28:05 crc kubenswrapper[4745]: I1010 13:28:05.148140 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" event={"ID":"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4","Type":"ContainerDied","Data":"8d9b26dd56eae4ae0f313726c1563a99fbf04fa2706a4e0d1875576f83a54ebd"} Oct 10 13:28:05 crc kubenswrapper[4745]: I1010 13:28:05.148191 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" event={"ID":"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4","Type":"ContainerStarted","Data":"98b6207c33caa39fe0cb6405cc14f2e6edcf700cbc87d56cc011fe547827f20c"} Oct 10 13:28:06 crc kubenswrapper[4745]: I1010 13:28:06.946403 4745 scope.go:117] "RemoveContainer" containerID="79aa5758abaacf3a8076f68e10cd35063c0bf828e1b01d66ee8a131b39af5210" Oct 10 13:28:06 crc kubenswrapper[4745]: I1010 13:28:06.969924 4745 scope.go:117] "RemoveContainer" containerID="668fb16e5543f24cd2b93f606627d511560b74d218f48fabe602ac3781ea471d" Oct 10 13:28:07 crc kubenswrapper[4745]: I1010 13:28:07.000089 4745 scope.go:117] "RemoveContainer" containerID="5ee1f2f51d33c5973ddf345d61d8fdc13e7ce1e4200ed108b14c2f5da882fec1" Oct 10 13:28:07 crc kubenswrapper[4745]: I1010 13:28:07.023333 4745 scope.go:117] "RemoveContainer" containerID="982fed50e3906cb02074d5d261a8839d2f429b2118b8cdfd5a3759b3bd9b90d9" Oct 10 13:28:07 crc kubenswrapper[4745]: I1010 13:28:07.051979 4745 scope.go:117] "RemoveContainer" containerID="6fd725491a3ac52f08fb6bc205680de98b2532452bc72c741b6bb2effe9668b6" Oct 10 13:28:07 crc kubenswrapper[4745]: I1010 13:28:07.069704 4745 scope.go:117] "RemoveContainer" containerID="ffaacc3f9ee18376c676b485649e50acb554a5cafafe737d65600ceefc32e3ce" Oct 10 13:28:07 crc kubenswrapper[4745]: I1010 13:28:07.094688 4745 scope.go:117] "RemoveContainer" containerID="c70473a533949ed11cbfd43506a1429f0f0d68267701276682d2cc39bcd087e9" Oct 10 13:28:07 crc kubenswrapper[4745]: I1010 13:28:07.121062 4745 scope.go:117] "RemoveContainer" containerID="f299a99cc8171745c01264eeb61d2227d5eb4998ff8140d298b60bf43973ba69" Oct 10 13:28:07 crc kubenswrapper[4745]: I1010 13:28:07.141612 4745 scope.go:117] "RemoveContainer" containerID="70bf205cef6ae92dbf67b4f1d8649411dda803f93997651fd812ffc1108b7f5f" Oct 10 13:28:07 crc kubenswrapper[4745]: I1010 13:28:07.163957 4745 generic.go:334] "Generic (PLEG): container finished" podID="4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4" containerID="00746a1f6060396843889b03f47a7763a02fdec7045490caef2d96b7abd195f6" exitCode=0 Oct 10 13:28:07 crc kubenswrapper[4745]: I1010 13:28:07.164132 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" event={"ID":"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4","Type":"ContainerDied","Data":"00746a1f6060396843889b03f47a7763a02fdec7045490caef2d96b7abd195f6"} Oct 10 13:28:08 crc kubenswrapper[4745]: I1010 13:28:08.179829 4745 generic.go:334] "Generic (PLEG): container finished" podID="4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4" containerID="48ac4583640335edc01f460cef07c828ede7909769e4569677e87c332296752e" exitCode=0 Oct 10 13:28:08 crc kubenswrapper[4745]: I1010 13:28:08.179872 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" event={"ID":"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4","Type":"ContainerDied","Data":"48ac4583640335edc01f460cef07c828ede7909769e4569677e87c332296752e"} Oct 10 13:28:09 crc kubenswrapper[4745]: I1010 13:28:09.534483 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" Oct 10 13:28:09 crc kubenswrapper[4745]: I1010 13:28:09.630534 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-bundle\") pod \"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4\" (UID: \"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4\") " Oct 10 13:28:09 crc kubenswrapper[4745]: I1010 13:28:09.630652 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98nw7\" (UniqueName: \"kubernetes.io/projected/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-kube-api-access-98nw7\") pod \"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4\" (UID: \"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4\") " Oct 10 13:28:09 crc kubenswrapper[4745]: I1010 13:28:09.630713 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-util\") pod \"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4\" (UID: \"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4\") " Oct 10 13:28:09 crc kubenswrapper[4745]: I1010 13:28:09.631456 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-bundle" (OuterVolumeSpecName: "bundle") pod "4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4" (UID: "4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:28:09 crc kubenswrapper[4745]: I1010 13:28:09.636868 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-kube-api-access-98nw7" (OuterVolumeSpecName: "kube-api-access-98nw7") pod "4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4" (UID: "4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4"). InnerVolumeSpecName "kube-api-access-98nw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:28:09 crc kubenswrapper[4745]: I1010 13:28:09.650195 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-util" (OuterVolumeSpecName: "util") pod "4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4" (UID: "4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:28:09 crc kubenswrapper[4745]: I1010 13:28:09.732259 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98nw7\" (UniqueName: \"kubernetes.io/projected/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-kube-api-access-98nw7\") on node \"crc\" DevicePath \"\"" Oct 10 13:28:09 crc kubenswrapper[4745]: I1010 13:28:09.732299 4745 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-util\") on node \"crc\" DevicePath \"\"" Oct 10 13:28:09 crc kubenswrapper[4745]: I1010 13:28:09.732311 4745 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:28:10 crc kubenswrapper[4745]: I1010 13:28:10.201850 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" event={"ID":"4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4","Type":"ContainerDied","Data":"98b6207c33caa39fe0cb6405cc14f2e6edcf700cbc87d56cc011fe547827f20c"} Oct 10 13:28:10 crc kubenswrapper[4745]: I1010 13:28:10.201906 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98b6207c33caa39fe0cb6405cc14f2e6edcf700cbc87d56cc011fe547827f20c" Oct 10 13:28:10 crc kubenswrapper[4745]: I1010 13:28:10.201997 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c" Oct 10 13:28:12 crc kubenswrapper[4745]: I1010 13:28:12.318884 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-8sw7s"] Oct 10 13:28:12 crc kubenswrapper[4745]: E1010 13:28:12.319306 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4" containerName="pull" Oct 10 13:28:12 crc kubenswrapper[4745]: I1010 13:28:12.319317 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4" containerName="pull" Oct 10 13:28:12 crc kubenswrapper[4745]: E1010 13:28:12.319332 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4" containerName="util" Oct 10 13:28:12 crc kubenswrapper[4745]: I1010 13:28:12.319353 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4" containerName="util" Oct 10 13:28:12 crc kubenswrapper[4745]: E1010 13:28:12.319362 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4" containerName="extract" Oct 10 13:28:12 crc kubenswrapper[4745]: I1010 13:28:12.319367 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4" containerName="extract" Oct 10 13:28:12 crc kubenswrapper[4745]: I1010 13:28:12.319464 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4" containerName="extract" Oct 10 13:28:12 crc kubenswrapper[4745]: I1010 13:28:12.319891 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-8sw7s" Oct 10 13:28:12 crc kubenswrapper[4745]: I1010 13:28:12.321385 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-swh6z" Oct 10 13:28:12 crc kubenswrapper[4745]: I1010 13:28:12.321827 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 10 13:28:12 crc kubenswrapper[4745]: I1010 13:28:12.321917 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 10 13:28:12 crc kubenswrapper[4745]: I1010 13:28:12.326805 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-8sw7s"] Oct 10 13:28:12 crc kubenswrapper[4745]: I1010 13:28:12.468859 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld2j7\" (UniqueName: \"kubernetes.io/projected/f5837847-c246-4661-9174-b0ec21039982-kube-api-access-ld2j7\") pod \"nmstate-operator-858ddd8f98-8sw7s\" (UID: \"f5837847-c246-4661-9174-b0ec21039982\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-8sw7s" Oct 10 13:28:12 crc kubenswrapper[4745]: I1010 13:28:12.570867 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld2j7\" (UniqueName: \"kubernetes.io/projected/f5837847-c246-4661-9174-b0ec21039982-kube-api-access-ld2j7\") pod \"nmstate-operator-858ddd8f98-8sw7s\" (UID: \"f5837847-c246-4661-9174-b0ec21039982\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-8sw7s" Oct 10 13:28:12 crc kubenswrapper[4745]: I1010 13:28:12.597336 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld2j7\" (UniqueName: \"kubernetes.io/projected/f5837847-c246-4661-9174-b0ec21039982-kube-api-access-ld2j7\") pod \"nmstate-operator-858ddd8f98-8sw7s\" (UID: \"f5837847-c246-4661-9174-b0ec21039982\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-8sw7s" Oct 10 13:28:12 crc kubenswrapper[4745]: I1010 13:28:12.632232 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-8sw7s" Oct 10 13:28:12 crc kubenswrapper[4745]: I1010 13:28:12.873194 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-8sw7s"] Oct 10 13:28:13 crc kubenswrapper[4745]: I1010 13:28:13.228696 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-8sw7s" event={"ID":"f5837847-c246-4661-9174-b0ec21039982","Type":"ContainerStarted","Data":"b3403032ce6ac2e3e5179e0fcdad2139daa241dd1fa2ff39dd6ce8e414d19007"} Oct 10 13:28:16 crc kubenswrapper[4745]: I1010 13:28:16.250381 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-8sw7s" event={"ID":"f5837847-c246-4661-9174-b0ec21039982","Type":"ContainerStarted","Data":"347459c922a4b00729eec490001891fc54fd3515c5b0a71fa21a76f6aa2f7407"} Oct 10 13:28:16 crc kubenswrapper[4745]: I1010 13:28:16.282542 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-8sw7s" podStartSLOduration=1.440596 podStartE2EDuration="4.282511286s" podCreationTimestamp="2025-10-10 13:28:12 +0000 UTC" firstStartedPulling="2025-10-10 13:28:12.880155142 +0000 UTC m=+606.777811925" lastFinishedPulling="2025-10-10 13:28:15.722070448 +0000 UTC m=+609.619727211" observedRunningTime="2025-10-10 13:28:16.275356547 +0000 UTC m=+610.173013350" watchObservedRunningTime="2025-10-10 13:28:16.282511286 +0000 UTC m=+610.180168089" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.229095 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-bsftc"] Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.231810 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bsftc" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.234069 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-fplpm" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.239564 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-s6hqt"] Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.240996 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-s6hqt" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.246374 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-bsftc"] Oct 10 13:28:17 crc kubenswrapper[4745]: W1010 13:28:17.248209 4745 reflector.go:561] object-"openshift-nmstate"/"openshift-nmstate-webhook": failed to list *v1.Secret: secrets "openshift-nmstate-webhook" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Oct 10 13:28:17 crc kubenswrapper[4745]: E1010 13:28:17.248343 4745 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"openshift-nmstate-webhook\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-nmstate-webhook\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.259908 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-s6hqt"] Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.279331 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-6mcw2"] Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.280312 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6mcw2" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.330978 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/869f5cad-555c-4695-a9bc-f9e699484561-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-s6hqt\" (UID: \"869f5cad-555c-4695-a9bc-f9e699484561\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-s6hqt" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.331123 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlk6b\" (UniqueName: \"kubernetes.io/projected/869f5cad-555c-4695-a9bc-f9e699484561-kube-api-access-mlk6b\") pod \"nmstate-webhook-6cdbc54649-s6hqt\" (UID: \"869f5cad-555c-4695-a9bc-f9e699484561\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-s6hqt" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.331191 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52vrb\" (UniqueName: \"kubernetes.io/projected/90c1fae7-ac4c-4d71-ac19-417d3f61f010-kube-api-access-52vrb\") pod \"nmstate-metrics-fdff9cb8d-bsftc\" (UID: \"90c1fae7-ac4c-4d71-ac19-417d3f61f010\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bsftc" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.378873 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-ws6j9"] Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.380059 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-ws6j9"] Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.380128 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ws6j9" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.382413 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-9j6c8" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.386147 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.386150 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.432942 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlk6b\" (UniqueName: \"kubernetes.io/projected/869f5cad-555c-4695-a9bc-f9e699484561-kube-api-access-mlk6b\") pod \"nmstate-webhook-6cdbc54649-s6hqt\" (UID: \"869f5cad-555c-4695-a9bc-f9e699484561\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-s6hqt" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.432988 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/9c18fb6e-a0ef-4367-a917-1abde904faf2-ovs-socket\") pod \"nmstate-handler-6mcw2\" (UID: \"9c18fb6e-a0ef-4367-a917-1abde904faf2\") " pod="openshift-nmstate/nmstate-handler-6mcw2" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.433031 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52vrb\" (UniqueName: \"kubernetes.io/projected/90c1fae7-ac4c-4d71-ac19-417d3f61f010-kube-api-access-52vrb\") pod \"nmstate-metrics-fdff9cb8d-bsftc\" (UID: \"90c1fae7-ac4c-4d71-ac19-417d3f61f010\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bsftc" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.433050 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85667\" (UniqueName: \"kubernetes.io/projected/9c18fb6e-a0ef-4367-a917-1abde904faf2-kube-api-access-85667\") pod \"nmstate-handler-6mcw2\" (UID: \"9c18fb6e-a0ef-4367-a917-1abde904faf2\") " pod="openshift-nmstate/nmstate-handler-6mcw2" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.433278 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/9c18fb6e-a0ef-4367-a917-1abde904faf2-dbus-socket\") pod \"nmstate-handler-6mcw2\" (UID: \"9c18fb6e-a0ef-4367-a917-1abde904faf2\") " pod="openshift-nmstate/nmstate-handler-6mcw2" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.433302 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/9c18fb6e-a0ef-4367-a917-1abde904faf2-nmstate-lock\") pod \"nmstate-handler-6mcw2\" (UID: \"9c18fb6e-a0ef-4367-a917-1abde904faf2\") " pod="openshift-nmstate/nmstate-handler-6mcw2" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.433328 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/869f5cad-555c-4695-a9bc-f9e699484561-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-s6hqt\" (UID: \"869f5cad-555c-4695-a9bc-f9e699484561\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-s6hqt" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.449477 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52vrb\" (UniqueName: \"kubernetes.io/projected/90c1fae7-ac4c-4d71-ac19-417d3f61f010-kube-api-access-52vrb\") pod \"nmstate-metrics-fdff9cb8d-bsftc\" (UID: \"90c1fae7-ac4c-4d71-ac19-417d3f61f010\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bsftc" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.449504 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlk6b\" (UniqueName: \"kubernetes.io/projected/869f5cad-555c-4695-a9bc-f9e699484561-kube-api-access-mlk6b\") pod \"nmstate-webhook-6cdbc54649-s6hqt\" (UID: \"869f5cad-555c-4695-a9bc-f9e699484561\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-s6hqt" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.521840 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7b7976b88c-m4kpn"] Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.522475 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.533968 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b8628d70-5b97-4c1c-b8c1-10e28a0439dd-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-ws6j9\" (UID: \"b8628d70-5b97-4c1c-b8c1-10e28a0439dd\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ws6j9" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.534080 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85667\" (UniqueName: \"kubernetes.io/projected/9c18fb6e-a0ef-4367-a917-1abde904faf2-kube-api-access-85667\") pod \"nmstate-handler-6mcw2\" (UID: \"9c18fb6e-a0ef-4367-a917-1abde904faf2\") " pod="openshift-nmstate/nmstate-handler-6mcw2" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.534140 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfrj2\" (UniqueName: \"kubernetes.io/projected/b8628d70-5b97-4c1c-b8c1-10e28a0439dd-kube-api-access-rfrj2\") pod \"nmstate-console-plugin-6b874cbd85-ws6j9\" (UID: \"b8628d70-5b97-4c1c-b8c1-10e28a0439dd\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ws6j9" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.534186 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/9c18fb6e-a0ef-4367-a917-1abde904faf2-dbus-socket\") pod \"nmstate-handler-6mcw2\" (UID: \"9c18fb6e-a0ef-4367-a917-1abde904faf2\") " pod="openshift-nmstate/nmstate-handler-6mcw2" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.534208 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/9c18fb6e-a0ef-4367-a917-1abde904faf2-nmstate-lock\") pod \"nmstate-handler-6mcw2\" (UID: \"9c18fb6e-a0ef-4367-a917-1abde904faf2\") " pod="openshift-nmstate/nmstate-handler-6mcw2" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.534252 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b8628d70-5b97-4c1c-b8c1-10e28a0439dd-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-ws6j9\" (UID: \"b8628d70-5b97-4c1c-b8c1-10e28a0439dd\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ws6j9" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.534308 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/9c18fb6e-a0ef-4367-a917-1abde904faf2-ovs-socket\") pod \"nmstate-handler-6mcw2\" (UID: \"9c18fb6e-a0ef-4367-a917-1abde904faf2\") " pod="openshift-nmstate/nmstate-handler-6mcw2" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.534369 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/9c18fb6e-a0ef-4367-a917-1abde904faf2-ovs-socket\") pod \"nmstate-handler-6mcw2\" (UID: \"9c18fb6e-a0ef-4367-a917-1abde904faf2\") " pod="openshift-nmstate/nmstate-handler-6mcw2" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.534435 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/9c18fb6e-a0ef-4367-a917-1abde904faf2-nmstate-lock\") pod \"nmstate-handler-6mcw2\" (UID: \"9c18fb6e-a0ef-4367-a917-1abde904faf2\") " pod="openshift-nmstate/nmstate-handler-6mcw2" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.534609 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/9c18fb6e-a0ef-4367-a917-1abde904faf2-dbus-socket\") pod \"nmstate-handler-6mcw2\" (UID: \"9c18fb6e-a0ef-4367-a917-1abde904faf2\") " pod="openshift-nmstate/nmstate-handler-6mcw2" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.535647 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7b7976b88c-m4kpn"] Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.545473 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bsftc" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.559621 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85667\" (UniqueName: \"kubernetes.io/projected/9c18fb6e-a0ef-4367-a917-1abde904faf2-kube-api-access-85667\") pod \"nmstate-handler-6mcw2\" (UID: \"9c18fb6e-a0ef-4367-a917-1abde904faf2\") " pod="openshift-nmstate/nmstate-handler-6mcw2" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.592504 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6mcw2" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.635345 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfrj2\" (UniqueName: \"kubernetes.io/projected/b8628d70-5b97-4c1c-b8c1-10e28a0439dd-kube-api-access-rfrj2\") pod \"nmstate-console-plugin-6b874cbd85-ws6j9\" (UID: \"b8628d70-5b97-4c1c-b8c1-10e28a0439dd\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ws6j9" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.635390 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/73a53508-3186-43f9-bc6d-db9aea490b75-console-serving-cert\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.635437 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/73a53508-3186-43f9-bc6d-db9aea490b75-service-ca\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.635491 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b8628d70-5b97-4c1c-b8c1-10e28a0439dd-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-ws6j9\" (UID: \"b8628d70-5b97-4c1c-b8c1-10e28a0439dd\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ws6j9" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.635520 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8tqx\" (UniqueName: \"kubernetes.io/projected/73a53508-3186-43f9-bc6d-db9aea490b75-kube-api-access-w8tqx\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.635534 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/73a53508-3186-43f9-bc6d-db9aea490b75-oauth-serving-cert\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.635578 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/73a53508-3186-43f9-bc6d-db9aea490b75-console-config\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.635597 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b8628d70-5b97-4c1c-b8c1-10e28a0439dd-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-ws6j9\" (UID: \"b8628d70-5b97-4c1c-b8c1-10e28a0439dd\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ws6j9" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.635618 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/73a53508-3186-43f9-bc6d-db9aea490b75-console-oauth-config\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.635652 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/73a53508-3186-43f9-bc6d-db9aea490b75-trusted-ca-bundle\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.638503 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b8628d70-5b97-4c1c-b8c1-10e28a0439dd-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-ws6j9\" (UID: \"b8628d70-5b97-4c1c-b8c1-10e28a0439dd\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ws6j9" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.641572 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b8628d70-5b97-4c1c-b8c1-10e28a0439dd-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-ws6j9\" (UID: \"b8628d70-5b97-4c1c-b8c1-10e28a0439dd\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ws6j9" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.657397 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfrj2\" (UniqueName: \"kubernetes.io/projected/b8628d70-5b97-4c1c-b8c1-10e28a0439dd-kube-api-access-rfrj2\") pod \"nmstate-console-plugin-6b874cbd85-ws6j9\" (UID: \"b8628d70-5b97-4c1c-b8c1-10e28a0439dd\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ws6j9" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.696629 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ws6j9" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.740390 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/73a53508-3186-43f9-bc6d-db9aea490b75-console-oauth-config\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.740439 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/73a53508-3186-43f9-bc6d-db9aea490b75-trusted-ca-bundle\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.740529 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/73a53508-3186-43f9-bc6d-db9aea490b75-console-serving-cert\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.740561 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/73a53508-3186-43f9-bc6d-db9aea490b75-service-ca\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.741059 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8tqx\" (UniqueName: \"kubernetes.io/projected/73a53508-3186-43f9-bc6d-db9aea490b75-kube-api-access-w8tqx\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.741100 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/73a53508-3186-43f9-bc6d-db9aea490b75-oauth-serving-cert\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.741137 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/73a53508-3186-43f9-bc6d-db9aea490b75-console-config\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.741703 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/73a53508-3186-43f9-bc6d-db9aea490b75-service-ca\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.742207 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/73a53508-3186-43f9-bc6d-db9aea490b75-console-config\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.742422 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/73a53508-3186-43f9-bc6d-db9aea490b75-oauth-serving-cert\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.743509 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/73a53508-3186-43f9-bc6d-db9aea490b75-trusted-ca-bundle\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.744368 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/73a53508-3186-43f9-bc6d-db9aea490b75-console-serving-cert\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.744668 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/73a53508-3186-43f9-bc6d-db9aea490b75-console-oauth-config\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.758646 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8tqx\" (UniqueName: \"kubernetes.io/projected/73a53508-3186-43f9-bc6d-db9aea490b75-kube-api-access-w8tqx\") pod \"console-7b7976b88c-m4kpn\" (UID: \"73a53508-3186-43f9-bc6d-db9aea490b75\") " pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.837556 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.847821 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-ws6j9"] Oct 10 13:28:17 crc kubenswrapper[4745]: I1010 13:28:17.935794 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-bsftc"] Oct 10 13:28:17 crc kubenswrapper[4745]: W1010 13:28:17.948410 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90c1fae7_ac4c_4d71_ac19_417d3f61f010.slice/crio-06a4c5c49a951794530437d8fd7555ce58789a88df816c4b7f30bc21d7a8ed00 WatchSource:0}: Error finding container 06a4c5c49a951794530437d8fd7555ce58789a88df816c4b7f30bc21d7a8ed00: Status 404 returned error can't find the container with id 06a4c5c49a951794530437d8fd7555ce58789a88df816c4b7f30bc21d7a8ed00 Oct 10 13:28:18 crc kubenswrapper[4745]: I1010 13:28:18.050342 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7b7976b88c-m4kpn"] Oct 10 13:28:18 crc kubenswrapper[4745]: W1010 13:28:18.054961 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73a53508_3186_43f9_bc6d_db9aea490b75.slice/crio-4990293b31405aa5f813b9482ced0d784fa84eac44812560c888af912a2c2898 WatchSource:0}: Error finding container 4990293b31405aa5f813b9482ced0d784fa84eac44812560c888af912a2c2898: Status 404 returned error can't find the container with id 4990293b31405aa5f813b9482ced0d784fa84eac44812560c888af912a2c2898 Oct 10 13:28:18 crc kubenswrapper[4745]: I1010 13:28:18.163147 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 10 13:28:18 crc kubenswrapper[4745]: I1010 13:28:18.170261 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/869f5cad-555c-4695-a9bc-f9e699484561-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-s6hqt\" (UID: \"869f5cad-555c-4695-a9bc-f9e699484561\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-s6hqt" Oct 10 13:28:18 crc kubenswrapper[4745]: I1010 13:28:18.264014 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bsftc" event={"ID":"90c1fae7-ac4c-4d71-ac19-417d3f61f010","Type":"ContainerStarted","Data":"06a4c5c49a951794530437d8fd7555ce58789a88df816c4b7f30bc21d7a8ed00"} Oct 10 13:28:18 crc kubenswrapper[4745]: I1010 13:28:18.265831 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7b7976b88c-m4kpn" event={"ID":"73a53508-3186-43f9-bc6d-db9aea490b75","Type":"ContainerStarted","Data":"9f1ccd228124afab15b86462fc50e651c0f60ac84c4245ec473d710c04735fa4"} Oct 10 13:28:18 crc kubenswrapper[4745]: I1010 13:28:18.265887 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7b7976b88c-m4kpn" event={"ID":"73a53508-3186-43f9-bc6d-db9aea490b75","Type":"ContainerStarted","Data":"4990293b31405aa5f813b9482ced0d784fa84eac44812560c888af912a2c2898"} Oct 10 13:28:18 crc kubenswrapper[4745]: I1010 13:28:18.267543 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6mcw2" event={"ID":"9c18fb6e-a0ef-4367-a917-1abde904faf2","Type":"ContainerStarted","Data":"71ad00c5afe064306416fc6ebf4baab5d08ae5be694eae87f6e9912888efb145"} Oct 10 13:28:18 crc kubenswrapper[4745]: I1010 13:28:18.269074 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ws6j9" event={"ID":"b8628d70-5b97-4c1c-b8c1-10e28a0439dd","Type":"ContainerStarted","Data":"81a8baea9c14ab38b9e0a668f82f31a37bd201edcff2a771ceba72f64420201c"} Oct 10 13:28:18 crc kubenswrapper[4745]: I1010 13:28:18.295999 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7b7976b88c-m4kpn" podStartSLOduration=1.2959799570000001 podStartE2EDuration="1.295979957s" podCreationTimestamp="2025-10-10 13:28:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:28:18.292640939 +0000 UTC m=+612.190297772" watchObservedRunningTime="2025-10-10 13:28:18.295979957 +0000 UTC m=+612.193636720" Oct 10 13:28:18 crc kubenswrapper[4745]: I1010 13:28:18.459277 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-s6hqt" Oct 10 13:28:18 crc kubenswrapper[4745]: I1010 13:28:18.660160 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-s6hqt"] Oct 10 13:28:19 crc kubenswrapper[4745]: I1010 13:28:19.279907 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-s6hqt" event={"ID":"869f5cad-555c-4695-a9bc-f9e699484561","Type":"ContainerStarted","Data":"d38c8f4ab149b259994a820476571c64f36e03941adaad5f8e62b0e8676a64bc"} Oct 10 13:28:21 crc kubenswrapper[4745]: I1010 13:28:21.292683 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bsftc" event={"ID":"90c1fae7-ac4c-4d71-ac19-417d3f61f010","Type":"ContainerStarted","Data":"68010e5fd29334019d294d037b14e8c8e5d8c809f90280452c203686243cda56"} Oct 10 13:28:21 crc kubenswrapper[4745]: I1010 13:28:21.295229 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-s6hqt" event={"ID":"869f5cad-555c-4695-a9bc-f9e699484561","Type":"ContainerStarted","Data":"925f18b67d1101aa590b675405562882cb4487f59723e3e588e0c92078468532"} Oct 10 13:28:21 crc kubenswrapper[4745]: I1010 13:28:21.295445 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-s6hqt" Oct 10 13:28:21 crc kubenswrapper[4745]: I1010 13:28:21.297856 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6mcw2" event={"ID":"9c18fb6e-a0ef-4367-a917-1abde904faf2","Type":"ContainerStarted","Data":"4f3f06436a81e68c8648d281aa49b6e19ed71c74c2b552a3f12ecb0613835cbd"} Oct 10 13:28:21 crc kubenswrapper[4745]: I1010 13:28:21.297952 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-6mcw2" Oct 10 13:28:21 crc kubenswrapper[4745]: I1010 13:28:21.299712 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ws6j9" event={"ID":"b8628d70-5b97-4c1c-b8c1-10e28a0439dd","Type":"ContainerStarted","Data":"73cadaa930d3e6d4936189528bb57c52025dc8a62f2ccd3f40ab96c3edd81dff"} Oct 10 13:28:21 crc kubenswrapper[4745]: I1010 13:28:21.331284 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-6mcw2" podStartSLOduration=1.047722521 podStartE2EDuration="4.331265121s" podCreationTimestamp="2025-10-10 13:28:17 +0000 UTC" firstStartedPulling="2025-10-10 13:28:17.624556733 +0000 UTC m=+611.522213496" lastFinishedPulling="2025-10-10 13:28:20.908099333 +0000 UTC m=+614.805756096" observedRunningTime="2025-10-10 13:28:21.330937633 +0000 UTC m=+615.228594406" watchObservedRunningTime="2025-10-10 13:28:21.331265121 +0000 UTC m=+615.228921884" Oct 10 13:28:21 crc kubenswrapper[4745]: I1010 13:28:21.331630 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-s6hqt" podStartSLOduration=2.0914634420000002 podStartE2EDuration="4.331623001s" podCreationTimestamp="2025-10-10 13:28:17 +0000 UTC" firstStartedPulling="2025-10-10 13:28:18.669424764 +0000 UTC m=+612.567081527" lastFinishedPulling="2025-10-10 13:28:20.909584323 +0000 UTC m=+614.807241086" observedRunningTime="2025-10-10 13:28:21.315529347 +0000 UTC m=+615.213186120" watchObservedRunningTime="2025-10-10 13:28:21.331623001 +0000 UTC m=+615.229279764" Oct 10 13:28:21 crc kubenswrapper[4745]: I1010 13:28:21.356615 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ws6j9" podStartSLOduration=1.3243004329999999 podStartE2EDuration="4.356594879s" podCreationTimestamp="2025-10-10 13:28:17 +0000 UTC" firstStartedPulling="2025-10-10 13:28:17.858252895 +0000 UTC m=+611.755909658" lastFinishedPulling="2025-10-10 13:28:20.890547341 +0000 UTC m=+614.788204104" observedRunningTime="2025-10-10 13:28:21.354371211 +0000 UTC m=+615.252027984" watchObservedRunningTime="2025-10-10 13:28:21.356594879 +0000 UTC m=+615.254251662" Oct 10 13:28:24 crc kubenswrapper[4745]: I1010 13:28:24.321363 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bsftc" event={"ID":"90c1fae7-ac4c-4d71-ac19-417d3f61f010","Type":"ContainerStarted","Data":"2a0e58e88638d137667c850da232a0ee9c51e806d19f1d1e318bfafa2c45b23b"} Oct 10 13:28:24 crc kubenswrapper[4745]: I1010 13:28:24.348141 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bsftc" podStartSLOduration=1.7587609880000001 podStartE2EDuration="7.34811283s" podCreationTimestamp="2025-10-10 13:28:17 +0000 UTC" firstStartedPulling="2025-10-10 13:28:17.960937212 +0000 UTC m=+611.858593985" lastFinishedPulling="2025-10-10 13:28:23.550289064 +0000 UTC m=+617.447945827" observedRunningTime="2025-10-10 13:28:24.345500281 +0000 UTC m=+618.243157044" watchObservedRunningTime="2025-10-10 13:28:24.34811283 +0000 UTC m=+618.245769633" Oct 10 13:28:27 crc kubenswrapper[4745]: I1010 13:28:27.617689 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-6mcw2" Oct 10 13:28:27 crc kubenswrapper[4745]: I1010 13:28:27.838943 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:27 crc kubenswrapper[4745]: I1010 13:28:27.839015 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:27 crc kubenswrapper[4745]: I1010 13:28:27.847273 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:28 crc kubenswrapper[4745]: I1010 13:28:28.360532 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7b7976b88c-m4kpn" Oct 10 13:28:28 crc kubenswrapper[4745]: I1010 13:28:28.473036 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-xq5mm"] Oct 10 13:28:38 crc kubenswrapper[4745]: I1010 13:28:38.467818 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-s6hqt" Oct 10 13:28:51 crc kubenswrapper[4745]: I1010 13:28:51.849970 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl"] Oct 10 13:28:51 crc kubenswrapper[4745]: I1010 13:28:51.853438 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" Oct 10 13:28:51 crc kubenswrapper[4745]: I1010 13:28:51.856599 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 10 13:28:51 crc kubenswrapper[4745]: I1010 13:28:51.857833 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl"] Oct 10 13:28:51 crc kubenswrapper[4745]: I1010 13:28:51.944523 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2050d745-31cf-44f8-8fb8-e1a4701c9799-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl\" (UID: \"2050d745-31cf-44f8-8fb8-e1a4701c9799\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" Oct 10 13:28:51 crc kubenswrapper[4745]: I1010 13:28:51.944956 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2050d745-31cf-44f8-8fb8-e1a4701c9799-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl\" (UID: \"2050d745-31cf-44f8-8fb8-e1a4701c9799\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" Oct 10 13:28:51 crc kubenswrapper[4745]: I1010 13:28:51.945008 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znq5r\" (UniqueName: \"kubernetes.io/projected/2050d745-31cf-44f8-8fb8-e1a4701c9799-kube-api-access-znq5r\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl\" (UID: \"2050d745-31cf-44f8-8fb8-e1a4701c9799\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" Oct 10 13:28:52 crc kubenswrapper[4745]: I1010 13:28:52.046079 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2050d745-31cf-44f8-8fb8-e1a4701c9799-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl\" (UID: \"2050d745-31cf-44f8-8fb8-e1a4701c9799\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" Oct 10 13:28:52 crc kubenswrapper[4745]: I1010 13:28:52.046139 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znq5r\" (UniqueName: \"kubernetes.io/projected/2050d745-31cf-44f8-8fb8-e1a4701c9799-kube-api-access-znq5r\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl\" (UID: \"2050d745-31cf-44f8-8fb8-e1a4701c9799\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" Oct 10 13:28:52 crc kubenswrapper[4745]: I1010 13:28:52.046199 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2050d745-31cf-44f8-8fb8-e1a4701c9799-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl\" (UID: \"2050d745-31cf-44f8-8fb8-e1a4701c9799\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" Oct 10 13:28:52 crc kubenswrapper[4745]: I1010 13:28:52.047150 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2050d745-31cf-44f8-8fb8-e1a4701c9799-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl\" (UID: \"2050d745-31cf-44f8-8fb8-e1a4701c9799\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" Oct 10 13:28:52 crc kubenswrapper[4745]: I1010 13:28:52.047593 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2050d745-31cf-44f8-8fb8-e1a4701c9799-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl\" (UID: \"2050d745-31cf-44f8-8fb8-e1a4701c9799\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" Oct 10 13:28:52 crc kubenswrapper[4745]: I1010 13:28:52.085375 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znq5r\" (UniqueName: \"kubernetes.io/projected/2050d745-31cf-44f8-8fb8-e1a4701c9799-kube-api-access-znq5r\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl\" (UID: \"2050d745-31cf-44f8-8fb8-e1a4701c9799\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" Oct 10 13:28:52 crc kubenswrapper[4745]: I1010 13:28:52.180744 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" Oct 10 13:28:52 crc kubenswrapper[4745]: I1010 13:28:52.469620 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl"] Oct 10 13:28:52 crc kubenswrapper[4745]: I1010 13:28:52.514222 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" event={"ID":"2050d745-31cf-44f8-8fb8-e1a4701c9799","Type":"ContainerStarted","Data":"d78256ebad60613a6dc47855784e7152817a9364e49197030e143326247e1bae"} Oct 10 13:28:53 crc kubenswrapper[4745]: I1010 13:28:53.524814 4745 generic.go:334] "Generic (PLEG): container finished" podID="2050d745-31cf-44f8-8fb8-e1a4701c9799" containerID="034b5a5a336208e77051e47a1a1b2efe9c05bde59acd0b788e9a026c998f5d3d" exitCode=0 Oct 10 13:28:53 crc kubenswrapper[4745]: I1010 13:28:53.524897 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" event={"ID":"2050d745-31cf-44f8-8fb8-e1a4701c9799","Type":"ContainerDied","Data":"034b5a5a336208e77051e47a1a1b2efe9c05bde59acd0b788e9a026c998f5d3d"} Oct 10 13:28:53 crc kubenswrapper[4745]: I1010 13:28:53.548923 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-xq5mm" podUID="e8bcad89-0f86-4a07-9c81-aa8cac743319" containerName="console" containerID="cri-o://2cd9251382eb8c522d4acff8f4f742b889ec9d7015baa4decb92ed2f9d4cf0c4" gracePeriod=15 Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.042358 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-xq5mm_e8bcad89-0f86-4a07-9c81-aa8cac743319/console/0.log" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.042938 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.102479 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-trusted-ca-bundle\") pod \"e8bcad89-0f86-4a07-9c81-aa8cac743319\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.102576 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-oauth-serving-cert\") pod \"e8bcad89-0f86-4a07-9c81-aa8cac743319\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.102616 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htmd2\" (UniqueName: \"kubernetes.io/projected/e8bcad89-0f86-4a07-9c81-aa8cac743319-kube-api-access-htmd2\") pod \"e8bcad89-0f86-4a07-9c81-aa8cac743319\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.102652 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-service-ca\") pod \"e8bcad89-0f86-4a07-9c81-aa8cac743319\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.102676 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-config\") pod \"e8bcad89-0f86-4a07-9c81-aa8cac743319\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.102720 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-serving-cert\") pod \"e8bcad89-0f86-4a07-9c81-aa8cac743319\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.102800 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-oauth-config\") pod \"e8bcad89-0f86-4a07-9c81-aa8cac743319\" (UID: \"e8bcad89-0f86-4a07-9c81-aa8cac743319\") " Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.104074 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "e8bcad89-0f86-4a07-9c81-aa8cac743319" (UID: "e8bcad89-0f86-4a07-9c81-aa8cac743319"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.104158 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-service-ca" (OuterVolumeSpecName: "service-ca") pod "e8bcad89-0f86-4a07-9c81-aa8cac743319" (UID: "e8bcad89-0f86-4a07-9c81-aa8cac743319"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.104308 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-config" (OuterVolumeSpecName: "console-config") pod "e8bcad89-0f86-4a07-9c81-aa8cac743319" (UID: "e8bcad89-0f86-4a07-9c81-aa8cac743319"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.104986 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "e8bcad89-0f86-4a07-9c81-aa8cac743319" (UID: "e8bcad89-0f86-4a07-9c81-aa8cac743319"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.109989 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "e8bcad89-0f86-4a07-9c81-aa8cac743319" (UID: "e8bcad89-0f86-4a07-9c81-aa8cac743319"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.110906 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8bcad89-0f86-4a07-9c81-aa8cac743319-kube-api-access-htmd2" (OuterVolumeSpecName: "kube-api-access-htmd2") pod "e8bcad89-0f86-4a07-9c81-aa8cac743319" (UID: "e8bcad89-0f86-4a07-9c81-aa8cac743319"). InnerVolumeSpecName "kube-api-access-htmd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.113361 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "e8bcad89-0f86-4a07-9c81-aa8cac743319" (UID: "e8bcad89-0f86-4a07-9c81-aa8cac743319"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.204484 4745 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.204538 4745 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.204560 4745 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.204578 4745 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e8bcad89-0f86-4a07-9c81-aa8cac743319-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.204601 4745 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.204628 4745 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e8bcad89-0f86-4a07-9c81-aa8cac743319-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.204652 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htmd2\" (UniqueName: \"kubernetes.io/projected/e8bcad89-0f86-4a07-9c81-aa8cac743319-kube-api-access-htmd2\") on node \"crc\" DevicePath \"\"" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.534553 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-xq5mm_e8bcad89-0f86-4a07-9c81-aa8cac743319/console/0.log" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.535778 4745 generic.go:334] "Generic (PLEG): container finished" podID="e8bcad89-0f86-4a07-9c81-aa8cac743319" containerID="2cd9251382eb8c522d4acff8f4f742b889ec9d7015baa4decb92ed2f9d4cf0c4" exitCode=2 Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.535963 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xq5mm" event={"ID":"e8bcad89-0f86-4a07-9c81-aa8cac743319","Type":"ContainerDied","Data":"2cd9251382eb8c522d4acff8f4f742b889ec9d7015baa4decb92ed2f9d4cf0c4"} Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.536107 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xq5mm" event={"ID":"e8bcad89-0f86-4a07-9c81-aa8cac743319","Type":"ContainerDied","Data":"89e93c8e1cadc332351e61a749998383794f241505d7ca3c02dc02a103dc2b12"} Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.536107 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xq5mm" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.536179 4745 scope.go:117] "RemoveContainer" containerID="2cd9251382eb8c522d4acff8f4f742b889ec9d7015baa4decb92ed2f9d4cf0c4" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.566833 4745 scope.go:117] "RemoveContainer" containerID="2cd9251382eb8c522d4acff8f4f742b889ec9d7015baa4decb92ed2f9d4cf0c4" Oct 10 13:28:54 crc kubenswrapper[4745]: E1010 13:28:54.567564 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cd9251382eb8c522d4acff8f4f742b889ec9d7015baa4decb92ed2f9d4cf0c4\": container with ID starting with 2cd9251382eb8c522d4acff8f4f742b889ec9d7015baa4decb92ed2f9d4cf0c4 not found: ID does not exist" containerID="2cd9251382eb8c522d4acff8f4f742b889ec9d7015baa4decb92ed2f9d4cf0c4" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.567630 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cd9251382eb8c522d4acff8f4f742b889ec9d7015baa4decb92ed2f9d4cf0c4"} err="failed to get container status \"2cd9251382eb8c522d4acff8f4f742b889ec9d7015baa4decb92ed2f9d4cf0c4\": rpc error: code = NotFound desc = could not find container \"2cd9251382eb8c522d4acff8f4f742b889ec9d7015baa4decb92ed2f9d4cf0c4\": container with ID starting with 2cd9251382eb8c522d4acff8f4f742b889ec9d7015baa4decb92ed2f9d4cf0c4 not found: ID does not exist" Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.587876 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-xq5mm"] Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.590453 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-xq5mm"] Oct 10 13:28:54 crc kubenswrapper[4745]: I1010 13:28:54.756981 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8bcad89-0f86-4a07-9c81-aa8cac743319" path="/var/lib/kubelet/pods/e8bcad89-0f86-4a07-9c81-aa8cac743319/volumes" Oct 10 13:28:55 crc kubenswrapper[4745]: I1010 13:28:55.550362 4745 generic.go:334] "Generic (PLEG): container finished" podID="2050d745-31cf-44f8-8fb8-e1a4701c9799" containerID="e76773d075b2210fdc612adc2df9e0a95cfdc6864fd473b099424303a2e582b1" exitCode=0 Oct 10 13:28:55 crc kubenswrapper[4745]: I1010 13:28:55.550445 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" event={"ID":"2050d745-31cf-44f8-8fb8-e1a4701c9799","Type":"ContainerDied","Data":"e76773d075b2210fdc612adc2df9e0a95cfdc6864fd473b099424303a2e582b1"} Oct 10 13:28:56 crc kubenswrapper[4745]: I1010 13:28:56.564540 4745 generic.go:334] "Generic (PLEG): container finished" podID="2050d745-31cf-44f8-8fb8-e1a4701c9799" containerID="976e96af67fb1c10b2c132add26cd6277834b2ac5b42209a6bd59638c2771ae4" exitCode=0 Oct 10 13:28:56 crc kubenswrapper[4745]: I1010 13:28:56.564620 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" event={"ID":"2050d745-31cf-44f8-8fb8-e1a4701c9799","Type":"ContainerDied","Data":"976e96af67fb1c10b2c132add26cd6277834b2ac5b42209a6bd59638c2771ae4"} Oct 10 13:28:57 crc kubenswrapper[4745]: I1010 13:28:57.887794 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" Oct 10 13:28:57 crc kubenswrapper[4745]: I1010 13:28:57.962711 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znq5r\" (UniqueName: \"kubernetes.io/projected/2050d745-31cf-44f8-8fb8-e1a4701c9799-kube-api-access-znq5r\") pod \"2050d745-31cf-44f8-8fb8-e1a4701c9799\" (UID: \"2050d745-31cf-44f8-8fb8-e1a4701c9799\") " Oct 10 13:28:57 crc kubenswrapper[4745]: I1010 13:28:57.962978 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2050d745-31cf-44f8-8fb8-e1a4701c9799-util\") pod \"2050d745-31cf-44f8-8fb8-e1a4701c9799\" (UID: \"2050d745-31cf-44f8-8fb8-e1a4701c9799\") " Oct 10 13:28:57 crc kubenswrapper[4745]: I1010 13:28:57.963044 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2050d745-31cf-44f8-8fb8-e1a4701c9799-bundle\") pod \"2050d745-31cf-44f8-8fb8-e1a4701c9799\" (UID: \"2050d745-31cf-44f8-8fb8-e1a4701c9799\") " Oct 10 13:28:57 crc kubenswrapper[4745]: I1010 13:28:57.964907 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2050d745-31cf-44f8-8fb8-e1a4701c9799-bundle" (OuterVolumeSpecName: "bundle") pod "2050d745-31cf-44f8-8fb8-e1a4701c9799" (UID: "2050d745-31cf-44f8-8fb8-e1a4701c9799"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:28:57 crc kubenswrapper[4745]: I1010 13:28:57.973036 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2050d745-31cf-44f8-8fb8-e1a4701c9799-kube-api-access-znq5r" (OuterVolumeSpecName: "kube-api-access-znq5r") pod "2050d745-31cf-44f8-8fb8-e1a4701c9799" (UID: "2050d745-31cf-44f8-8fb8-e1a4701c9799"). InnerVolumeSpecName "kube-api-access-znq5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:28:58 crc kubenswrapper[4745]: I1010 13:28:58.065647 4745 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2050d745-31cf-44f8-8fb8-e1a4701c9799-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:28:58 crc kubenswrapper[4745]: I1010 13:28:58.065708 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znq5r\" (UniqueName: \"kubernetes.io/projected/2050d745-31cf-44f8-8fb8-e1a4701c9799-kube-api-access-znq5r\") on node \"crc\" DevicePath \"\"" Oct 10 13:28:58 crc kubenswrapper[4745]: I1010 13:28:58.341260 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2050d745-31cf-44f8-8fb8-e1a4701c9799-util" (OuterVolumeSpecName: "util") pod "2050d745-31cf-44f8-8fb8-e1a4701c9799" (UID: "2050d745-31cf-44f8-8fb8-e1a4701c9799"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:28:58 crc kubenswrapper[4745]: I1010 13:28:58.376781 4745 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2050d745-31cf-44f8-8fb8-e1a4701c9799-util\") on node \"crc\" DevicePath \"\"" Oct 10 13:28:58 crc kubenswrapper[4745]: I1010 13:28:58.586360 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" event={"ID":"2050d745-31cf-44f8-8fb8-e1a4701c9799","Type":"ContainerDied","Data":"d78256ebad60613a6dc47855784e7152817a9364e49197030e143326247e1bae"} Oct 10 13:28:58 crc kubenswrapper[4745]: I1010 13:28:58.586710 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d78256ebad60613a6dc47855784e7152817a9364e49197030e143326247e1bae" Oct 10 13:28:58 crc kubenswrapper[4745]: I1010 13:28:58.586451 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl" Oct 10 13:29:07 crc kubenswrapper[4745]: I1010 13:29:07.972276 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q"] Oct 10 13:29:07 crc kubenswrapper[4745]: E1010 13:29:07.972906 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8bcad89-0f86-4a07-9c81-aa8cac743319" containerName="console" Oct 10 13:29:07 crc kubenswrapper[4745]: I1010 13:29:07.972918 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8bcad89-0f86-4a07-9c81-aa8cac743319" containerName="console" Oct 10 13:29:07 crc kubenswrapper[4745]: E1010 13:29:07.972931 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2050d745-31cf-44f8-8fb8-e1a4701c9799" containerName="extract" Oct 10 13:29:07 crc kubenswrapper[4745]: I1010 13:29:07.972936 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="2050d745-31cf-44f8-8fb8-e1a4701c9799" containerName="extract" Oct 10 13:29:07 crc kubenswrapper[4745]: E1010 13:29:07.972949 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2050d745-31cf-44f8-8fb8-e1a4701c9799" containerName="pull" Oct 10 13:29:07 crc kubenswrapper[4745]: I1010 13:29:07.972955 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="2050d745-31cf-44f8-8fb8-e1a4701c9799" containerName="pull" Oct 10 13:29:07 crc kubenswrapper[4745]: E1010 13:29:07.972964 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2050d745-31cf-44f8-8fb8-e1a4701c9799" containerName="util" Oct 10 13:29:07 crc kubenswrapper[4745]: I1010 13:29:07.972971 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="2050d745-31cf-44f8-8fb8-e1a4701c9799" containerName="util" Oct 10 13:29:07 crc kubenswrapper[4745]: I1010 13:29:07.973063 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8bcad89-0f86-4a07-9c81-aa8cac743319" containerName="console" Oct 10 13:29:07 crc kubenswrapper[4745]: I1010 13:29:07.973075 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="2050d745-31cf-44f8-8fb8-e1a4701c9799" containerName="extract" Oct 10 13:29:07 crc kubenswrapper[4745]: I1010 13:29:07.973393 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q" Oct 10 13:29:07 crc kubenswrapper[4745]: I1010 13:29:07.978431 4745 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 10 13:29:07 crc kubenswrapper[4745]: I1010 13:29:07.978912 4745 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 10 13:29:07 crc kubenswrapper[4745]: I1010 13:29:07.978986 4745 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-wwf9r" Oct 10 13:29:07 crc kubenswrapper[4745]: I1010 13:29:07.979052 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 10 13:29:07 crc kubenswrapper[4745]: I1010 13:29:07.979330 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 10 13:29:07 crc kubenswrapper[4745]: I1010 13:29:07.991413 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q"] Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.103878 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xfjf\" (UniqueName: \"kubernetes.io/projected/bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1-kube-api-access-9xfjf\") pod \"metallb-operator-controller-manager-6d67b47897-wgz9q\" (UID: \"bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1\") " pod="metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.103945 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1-webhook-cert\") pod \"metallb-operator-controller-manager-6d67b47897-wgz9q\" (UID: \"bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1\") " pod="metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.103999 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1-apiservice-cert\") pod \"metallb-operator-controller-manager-6d67b47897-wgz9q\" (UID: \"bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1\") " pod="metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.205163 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1-webhook-cert\") pod \"metallb-operator-controller-manager-6d67b47897-wgz9q\" (UID: \"bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1\") " pod="metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.205239 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1-apiservice-cert\") pod \"metallb-operator-controller-manager-6d67b47897-wgz9q\" (UID: \"bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1\") " pod="metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.205270 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xfjf\" (UniqueName: \"kubernetes.io/projected/bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1-kube-api-access-9xfjf\") pod \"metallb-operator-controller-manager-6d67b47897-wgz9q\" (UID: \"bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1\") " pod="metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.213641 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1-apiservice-cert\") pod \"metallb-operator-controller-manager-6d67b47897-wgz9q\" (UID: \"bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1\") " pod="metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.226288 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1-webhook-cert\") pod \"metallb-operator-controller-manager-6d67b47897-wgz9q\" (UID: \"bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1\") " pod="metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.230525 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xfjf\" (UniqueName: \"kubernetes.io/projected/bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1-kube-api-access-9xfjf\") pod \"metallb-operator-controller-manager-6d67b47897-wgz9q\" (UID: \"bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1\") " pod="metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.288281 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.353591 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw"] Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.354328 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.359169 4745 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.359338 4745 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-r42lk" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.359445 4745 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.373774 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw"] Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.410789 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fcb7295e-30a8-4023-9f78-c75a0dadea04-apiservice-cert\") pod \"metallb-operator-webhook-server-55796c598f-6xdbw\" (UID: \"fcb7295e-30a8-4023-9f78-c75a0dadea04\") " pod="metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.410827 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk8np\" (UniqueName: \"kubernetes.io/projected/fcb7295e-30a8-4023-9f78-c75a0dadea04-kube-api-access-kk8np\") pod \"metallb-operator-webhook-server-55796c598f-6xdbw\" (UID: \"fcb7295e-30a8-4023-9f78-c75a0dadea04\") " pod="metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.410850 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fcb7295e-30a8-4023-9f78-c75a0dadea04-webhook-cert\") pod \"metallb-operator-webhook-server-55796c598f-6xdbw\" (UID: \"fcb7295e-30a8-4023-9f78-c75a0dadea04\") " pod="metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.512297 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fcb7295e-30a8-4023-9f78-c75a0dadea04-apiservice-cert\") pod \"metallb-operator-webhook-server-55796c598f-6xdbw\" (UID: \"fcb7295e-30a8-4023-9f78-c75a0dadea04\") " pod="metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.512338 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk8np\" (UniqueName: \"kubernetes.io/projected/fcb7295e-30a8-4023-9f78-c75a0dadea04-kube-api-access-kk8np\") pod \"metallb-operator-webhook-server-55796c598f-6xdbw\" (UID: \"fcb7295e-30a8-4023-9f78-c75a0dadea04\") " pod="metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.512370 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fcb7295e-30a8-4023-9f78-c75a0dadea04-webhook-cert\") pod \"metallb-operator-webhook-server-55796c598f-6xdbw\" (UID: \"fcb7295e-30a8-4023-9f78-c75a0dadea04\") " pod="metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.517586 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fcb7295e-30a8-4023-9f78-c75a0dadea04-webhook-cert\") pod \"metallb-operator-webhook-server-55796c598f-6xdbw\" (UID: \"fcb7295e-30a8-4023-9f78-c75a0dadea04\") " pod="metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.519300 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fcb7295e-30a8-4023-9f78-c75a0dadea04-apiservice-cert\") pod \"metallb-operator-webhook-server-55796c598f-6xdbw\" (UID: \"fcb7295e-30a8-4023-9f78-c75a0dadea04\") " pod="metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.538871 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk8np\" (UniqueName: \"kubernetes.io/projected/fcb7295e-30a8-4023-9f78-c75a0dadea04-kube-api-access-kk8np\") pod \"metallb-operator-webhook-server-55796c598f-6xdbw\" (UID: \"fcb7295e-30a8-4023-9f78-c75a0dadea04\") " pod="metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.680662 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw" Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.856083 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw"] Oct 10 13:29:08 crc kubenswrapper[4745]: W1010 13:29:08.865619 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfcb7295e_30a8_4023_9f78_c75a0dadea04.slice/crio-67ba1c00a79fd29cc602bc038cfc78f873d2079c0fa8398d7e1e431d0452c347 WatchSource:0}: Error finding container 67ba1c00a79fd29cc602bc038cfc78f873d2079c0fa8398d7e1e431d0452c347: Status 404 returned error can't find the container with id 67ba1c00a79fd29cc602bc038cfc78f873d2079c0fa8398d7e1e431d0452c347 Oct 10 13:29:08 crc kubenswrapper[4745]: I1010 13:29:08.876932 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q"] Oct 10 13:29:08 crc kubenswrapper[4745]: W1010 13:29:08.884042 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbdf86da2_3d6e_44f6_bd60_1ddfad7c6ff1.slice/crio-7d07351f1fb64ee53832f9df31075aea597f18e8a28b5c18ad7bada75d0b98a1 WatchSource:0}: Error finding container 7d07351f1fb64ee53832f9df31075aea597f18e8a28b5c18ad7bada75d0b98a1: Status 404 returned error can't find the container with id 7d07351f1fb64ee53832f9df31075aea597f18e8a28b5c18ad7bada75d0b98a1 Oct 10 13:29:09 crc kubenswrapper[4745]: I1010 13:29:09.646715 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw" event={"ID":"fcb7295e-30a8-4023-9f78-c75a0dadea04","Type":"ContainerStarted","Data":"67ba1c00a79fd29cc602bc038cfc78f873d2079c0fa8398d7e1e431d0452c347"} Oct 10 13:29:09 crc kubenswrapper[4745]: I1010 13:29:09.648223 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q" event={"ID":"bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1","Type":"ContainerStarted","Data":"7d07351f1fb64ee53832f9df31075aea597f18e8a28b5c18ad7bada75d0b98a1"} Oct 10 13:29:14 crc kubenswrapper[4745]: I1010 13:29:14.697621 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q" event={"ID":"bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1","Type":"ContainerStarted","Data":"b9db168d1f80669563294b4c83a6d3890dc089ee7d611bece61d9a6ad68ac71d"} Oct 10 13:29:14 crc kubenswrapper[4745]: I1010 13:29:14.700645 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw" event={"ID":"fcb7295e-30a8-4023-9f78-c75a0dadea04","Type":"ContainerStarted","Data":"0a2688714c8ab9967a08bd8af6e8533825cdbdf9f5c83aab4a5862efdb3c4c9c"} Oct 10 13:29:14 crc kubenswrapper[4745]: I1010 13:29:14.700874 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw" Oct 10 13:29:14 crc kubenswrapper[4745]: I1010 13:29:14.701377 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q" Oct 10 13:29:14 crc kubenswrapper[4745]: I1010 13:29:14.737160 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw" podStartSLOduration=1.970058428 podStartE2EDuration="6.737125354s" podCreationTimestamp="2025-10-10 13:29:08 +0000 UTC" firstStartedPulling="2025-10-10 13:29:08.872242561 +0000 UTC m=+662.769899324" lastFinishedPulling="2025-10-10 13:29:13.639309477 +0000 UTC m=+667.536966250" observedRunningTime="2025-10-10 13:29:14.729569958 +0000 UTC m=+668.627226751" watchObservedRunningTime="2025-10-10 13:29:14.737125354 +0000 UTC m=+668.634782157" Oct 10 13:29:14 crc kubenswrapper[4745]: I1010 13:29:14.763390 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q" podStartSLOduration=3.039918003 podStartE2EDuration="7.763363994s" podCreationTimestamp="2025-10-10 13:29:07 +0000 UTC" firstStartedPulling="2025-10-10 13:29:08.886725258 +0000 UTC m=+662.784382031" lastFinishedPulling="2025-10-10 13:29:13.610171249 +0000 UTC m=+667.507828022" observedRunningTime="2025-10-10 13:29:14.762899454 +0000 UTC m=+668.660556247" watchObservedRunningTime="2025-10-10 13:29:14.763363994 +0000 UTC m=+668.661020797" Oct 10 13:29:28 crc kubenswrapper[4745]: I1010 13:29:28.694213 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-55796c598f-6xdbw" Oct 10 13:29:46 crc kubenswrapper[4745]: I1010 13:29:46.186703 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:29:46 crc kubenswrapper[4745]: I1010 13:29:46.187384 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:29:48 crc kubenswrapper[4745]: I1010 13:29:48.291198 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6d67b47897-wgz9q" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.198218 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-8ktnf"] Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.199430 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8ktnf" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.201441 4745 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-csdhq" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.201755 4745 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.209019 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-52wk2"] Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.210985 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.212852 4745 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.212973 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.225036 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-8ktnf"] Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.262949 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e3feebb0-e7f8-4aef-a725-2b9cf003364c-frr-conf\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.262989 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3feebb0-e7f8-4aef-a725-2b9cf003364c-metrics-certs\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.263010 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a362b7b7-7534-4a4f-ae57-aac84fd292c7-cert\") pod \"frr-k8s-webhook-server-64bf5d555-8ktnf\" (UID: \"a362b7b7-7534-4a4f-ae57-aac84fd292c7\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8ktnf" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.263028 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e3feebb0-e7f8-4aef-a725-2b9cf003364c-frr-sockets\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.263047 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e3feebb0-e7f8-4aef-a725-2b9cf003364c-reloader\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.263310 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwftk\" (UniqueName: \"kubernetes.io/projected/e3feebb0-e7f8-4aef-a725-2b9cf003364c-kube-api-access-nwftk\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.263360 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e3feebb0-e7f8-4aef-a725-2b9cf003364c-frr-startup\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.263481 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e3feebb0-e7f8-4aef-a725-2b9cf003364c-metrics\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.263507 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdflr\" (UniqueName: \"kubernetes.io/projected/a362b7b7-7534-4a4f-ae57-aac84fd292c7-kube-api-access-rdflr\") pod \"frr-k8s-webhook-server-64bf5d555-8ktnf\" (UID: \"a362b7b7-7534-4a4f-ae57-aac84fd292c7\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8ktnf" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.304765 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-92z9w"] Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.305566 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-92z9w" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.308487 4745 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-slv6b" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.308672 4745 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.311367 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.312770 4745 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.317602 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-ktgdl"] Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.319006 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-ktgdl" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.320258 4745 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.338235 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-ktgdl"] Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.364385 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e3feebb0-e7f8-4aef-a725-2b9cf003364c-frr-sockets\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.364431 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e3feebb0-e7f8-4aef-a725-2b9cf003364c-reloader\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.364452 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b298d9ff-9b10-42f4-aabf-259673933a1a-metallb-excludel2\") pod \"speaker-92z9w\" (UID: \"b298d9ff-9b10-42f4-aabf-259673933a1a\") " pod="metallb-system/speaker-92z9w" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.364477 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c1711da0-4710-4569-bdbd-2afe523b91be-cert\") pod \"controller-68d546b9d8-ktgdl\" (UID: \"c1711da0-4710-4569-bdbd-2afe523b91be\") " pod="metallb-system/controller-68d546b9d8-ktgdl" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.364492 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s47rx\" (UniqueName: \"kubernetes.io/projected/b298d9ff-9b10-42f4-aabf-259673933a1a-kube-api-access-s47rx\") pod \"speaker-92z9w\" (UID: \"b298d9ff-9b10-42f4-aabf-259673933a1a\") " pod="metallb-system/speaker-92z9w" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.364516 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1711da0-4710-4569-bdbd-2afe523b91be-metrics-certs\") pod \"controller-68d546b9d8-ktgdl\" (UID: \"c1711da0-4710-4569-bdbd-2afe523b91be\") " pod="metallb-system/controller-68d546b9d8-ktgdl" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.364538 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwftk\" (UniqueName: \"kubernetes.io/projected/e3feebb0-e7f8-4aef-a725-2b9cf003364c-kube-api-access-nwftk\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.364554 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b298d9ff-9b10-42f4-aabf-259673933a1a-metrics-certs\") pod \"speaker-92z9w\" (UID: \"b298d9ff-9b10-42f4-aabf-259673933a1a\") " pod="metallb-system/speaker-92z9w" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.364569 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e3feebb0-e7f8-4aef-a725-2b9cf003364c-frr-startup\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.364597 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e3feebb0-e7f8-4aef-a725-2b9cf003364c-metrics\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.364612 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdflr\" (UniqueName: \"kubernetes.io/projected/a362b7b7-7534-4a4f-ae57-aac84fd292c7-kube-api-access-rdflr\") pod \"frr-k8s-webhook-server-64bf5d555-8ktnf\" (UID: \"a362b7b7-7534-4a4f-ae57-aac84fd292c7\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8ktnf" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.364629 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct6dh\" (UniqueName: \"kubernetes.io/projected/c1711da0-4710-4569-bdbd-2afe523b91be-kube-api-access-ct6dh\") pod \"controller-68d546b9d8-ktgdl\" (UID: \"c1711da0-4710-4569-bdbd-2afe523b91be\") " pod="metallb-system/controller-68d546b9d8-ktgdl" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.364649 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e3feebb0-e7f8-4aef-a725-2b9cf003364c-frr-conf\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.364686 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3feebb0-e7f8-4aef-a725-2b9cf003364c-metrics-certs\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.364702 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b298d9ff-9b10-42f4-aabf-259673933a1a-memberlist\") pod \"speaker-92z9w\" (UID: \"b298d9ff-9b10-42f4-aabf-259673933a1a\") " pod="metallb-system/speaker-92z9w" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.364717 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a362b7b7-7534-4a4f-ae57-aac84fd292c7-cert\") pod \"frr-k8s-webhook-server-64bf5d555-8ktnf\" (UID: \"a362b7b7-7534-4a4f-ae57-aac84fd292c7\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8ktnf" Oct 10 13:29:49 crc kubenswrapper[4745]: E1010 13:29:49.364821 4745 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 10 13:29:49 crc kubenswrapper[4745]: E1010 13:29:49.364866 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a362b7b7-7534-4a4f-ae57-aac84fd292c7-cert podName:a362b7b7-7534-4a4f-ae57-aac84fd292c7 nodeName:}" failed. No retries permitted until 2025-10-10 13:29:49.864851923 +0000 UTC m=+703.762508686 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a362b7b7-7534-4a4f-ae57-aac84fd292c7-cert") pod "frr-k8s-webhook-server-64bf5d555-8ktnf" (UID: "a362b7b7-7534-4a4f-ae57-aac84fd292c7") : secret "frr-k8s-webhook-server-cert" not found Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.365181 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e3feebb0-e7f8-4aef-a725-2b9cf003364c-frr-sockets\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.365350 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e3feebb0-e7f8-4aef-a725-2b9cf003364c-reloader\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: E1010 13:29:49.366000 4745 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 10 13:29:49 crc kubenswrapper[4745]: E1010 13:29:49.366083 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3feebb0-e7f8-4aef-a725-2b9cf003364c-metrics-certs podName:e3feebb0-e7f8-4aef-a725-2b9cf003364c nodeName:}" failed. No retries permitted until 2025-10-10 13:29:49.866057391 +0000 UTC m=+703.763714174 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e3feebb0-e7f8-4aef-a725-2b9cf003364c-metrics-certs") pod "frr-k8s-52wk2" (UID: "e3feebb0-e7f8-4aef-a725-2b9cf003364c") : secret "frr-k8s-certs-secret" not found Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.366130 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e3feebb0-e7f8-4aef-a725-2b9cf003364c-frr-conf\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.366349 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e3feebb0-e7f8-4aef-a725-2b9cf003364c-frr-startup\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.366352 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e3feebb0-e7f8-4aef-a725-2b9cf003364c-metrics\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.389566 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwftk\" (UniqueName: \"kubernetes.io/projected/e3feebb0-e7f8-4aef-a725-2b9cf003364c-kube-api-access-nwftk\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.390319 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdflr\" (UniqueName: \"kubernetes.io/projected/a362b7b7-7534-4a4f-ae57-aac84fd292c7-kube-api-access-rdflr\") pod \"frr-k8s-webhook-server-64bf5d555-8ktnf\" (UID: \"a362b7b7-7534-4a4f-ae57-aac84fd292c7\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8ktnf" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.466306 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b298d9ff-9b10-42f4-aabf-259673933a1a-metallb-excludel2\") pod \"speaker-92z9w\" (UID: \"b298d9ff-9b10-42f4-aabf-259673933a1a\") " pod="metallb-system/speaker-92z9w" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.466360 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c1711da0-4710-4569-bdbd-2afe523b91be-cert\") pod \"controller-68d546b9d8-ktgdl\" (UID: \"c1711da0-4710-4569-bdbd-2afe523b91be\") " pod="metallb-system/controller-68d546b9d8-ktgdl" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.466383 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s47rx\" (UniqueName: \"kubernetes.io/projected/b298d9ff-9b10-42f4-aabf-259673933a1a-kube-api-access-s47rx\") pod \"speaker-92z9w\" (UID: \"b298d9ff-9b10-42f4-aabf-259673933a1a\") " pod="metallb-system/speaker-92z9w" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.466411 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1711da0-4710-4569-bdbd-2afe523b91be-metrics-certs\") pod \"controller-68d546b9d8-ktgdl\" (UID: \"c1711da0-4710-4569-bdbd-2afe523b91be\") " pod="metallb-system/controller-68d546b9d8-ktgdl" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.466439 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b298d9ff-9b10-42f4-aabf-259673933a1a-metrics-certs\") pod \"speaker-92z9w\" (UID: \"b298d9ff-9b10-42f4-aabf-259673933a1a\") " pod="metallb-system/speaker-92z9w" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.466477 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct6dh\" (UniqueName: \"kubernetes.io/projected/c1711da0-4710-4569-bdbd-2afe523b91be-kube-api-access-ct6dh\") pod \"controller-68d546b9d8-ktgdl\" (UID: \"c1711da0-4710-4569-bdbd-2afe523b91be\") " pod="metallb-system/controller-68d546b9d8-ktgdl" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.466512 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b298d9ff-9b10-42f4-aabf-259673933a1a-memberlist\") pod \"speaker-92z9w\" (UID: \"b298d9ff-9b10-42f4-aabf-259673933a1a\") " pod="metallb-system/speaker-92z9w" Oct 10 13:29:49 crc kubenswrapper[4745]: E1010 13:29:49.466615 4745 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 10 13:29:49 crc kubenswrapper[4745]: E1010 13:29:49.466662 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b298d9ff-9b10-42f4-aabf-259673933a1a-memberlist podName:b298d9ff-9b10-42f4-aabf-259673933a1a nodeName:}" failed. No retries permitted until 2025-10-10 13:29:49.966649082 +0000 UTC m=+703.864305845 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b298d9ff-9b10-42f4-aabf-259673933a1a-memberlist") pod "speaker-92z9w" (UID: "b298d9ff-9b10-42f4-aabf-259673933a1a") : secret "metallb-memberlist" not found Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.467010 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b298d9ff-9b10-42f4-aabf-259673933a1a-metallb-excludel2\") pod \"speaker-92z9w\" (UID: \"b298d9ff-9b10-42f4-aabf-259673933a1a\") " pod="metallb-system/speaker-92z9w" Oct 10 13:29:49 crc kubenswrapper[4745]: E1010 13:29:49.467244 4745 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 10 13:29:49 crc kubenswrapper[4745]: E1010 13:29:49.467274 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b298d9ff-9b10-42f4-aabf-259673933a1a-metrics-certs podName:b298d9ff-9b10-42f4-aabf-259673933a1a nodeName:}" failed. No retries permitted until 2025-10-10 13:29:49.967265436 +0000 UTC m=+703.864922199 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b298d9ff-9b10-42f4-aabf-259673933a1a-metrics-certs") pod "speaker-92z9w" (UID: "b298d9ff-9b10-42f4-aabf-259673933a1a") : secret "speaker-certs-secret" not found Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.469783 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1711da0-4710-4569-bdbd-2afe523b91be-metrics-certs\") pod \"controller-68d546b9d8-ktgdl\" (UID: \"c1711da0-4710-4569-bdbd-2afe523b91be\") " pod="metallb-system/controller-68d546b9d8-ktgdl" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.470067 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c1711da0-4710-4569-bdbd-2afe523b91be-cert\") pod \"controller-68d546b9d8-ktgdl\" (UID: \"c1711da0-4710-4569-bdbd-2afe523b91be\") " pod="metallb-system/controller-68d546b9d8-ktgdl" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.489814 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct6dh\" (UniqueName: \"kubernetes.io/projected/c1711da0-4710-4569-bdbd-2afe523b91be-kube-api-access-ct6dh\") pod \"controller-68d546b9d8-ktgdl\" (UID: \"c1711da0-4710-4569-bdbd-2afe523b91be\") " pod="metallb-system/controller-68d546b9d8-ktgdl" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.492450 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s47rx\" (UniqueName: \"kubernetes.io/projected/b298d9ff-9b10-42f4-aabf-259673933a1a-kube-api-access-s47rx\") pod \"speaker-92z9w\" (UID: \"b298d9ff-9b10-42f4-aabf-259673933a1a\") " pod="metallb-system/speaker-92z9w" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.632102 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-ktgdl" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.874403 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3feebb0-e7f8-4aef-a725-2b9cf003364c-metrics-certs\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.874450 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a362b7b7-7534-4a4f-ae57-aac84fd292c7-cert\") pod \"frr-k8s-webhook-server-64bf5d555-8ktnf\" (UID: \"a362b7b7-7534-4a4f-ae57-aac84fd292c7\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8ktnf" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.876755 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-ktgdl"] Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.878482 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a362b7b7-7534-4a4f-ae57-aac84fd292c7-cert\") pod \"frr-k8s-webhook-server-64bf5d555-8ktnf\" (UID: \"a362b7b7-7534-4a4f-ae57-aac84fd292c7\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8ktnf" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.879884 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3feebb0-e7f8-4aef-a725-2b9cf003364c-metrics-certs\") pod \"frr-k8s-52wk2\" (UID: \"e3feebb0-e7f8-4aef-a725-2b9cf003364c\") " pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:49 crc kubenswrapper[4745]: W1010 13:29:49.880809 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1711da0_4710_4569_bdbd_2afe523b91be.slice/crio-c9ba34d10413620678e2fcfe04b39ed2d79782397bd41e2aa736ac2214ba03bc WatchSource:0}: Error finding container c9ba34d10413620678e2fcfe04b39ed2d79782397bd41e2aa736ac2214ba03bc: Status 404 returned error can't find the container with id c9ba34d10413620678e2fcfe04b39ed2d79782397bd41e2aa736ac2214ba03bc Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.928629 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-ktgdl" event={"ID":"c1711da0-4710-4569-bdbd-2afe523b91be","Type":"ContainerStarted","Data":"c9ba34d10413620678e2fcfe04b39ed2d79782397bd41e2aa736ac2214ba03bc"} Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.976321 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b298d9ff-9b10-42f4-aabf-259673933a1a-metrics-certs\") pod \"speaker-92z9w\" (UID: \"b298d9ff-9b10-42f4-aabf-259673933a1a\") " pod="metallb-system/speaker-92z9w" Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.976451 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b298d9ff-9b10-42f4-aabf-259673933a1a-memberlist\") pod \"speaker-92z9w\" (UID: \"b298d9ff-9b10-42f4-aabf-259673933a1a\") " pod="metallb-system/speaker-92z9w" Oct 10 13:29:49 crc kubenswrapper[4745]: E1010 13:29:49.976602 4745 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 10 13:29:49 crc kubenswrapper[4745]: E1010 13:29:49.976692 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b298d9ff-9b10-42f4-aabf-259673933a1a-memberlist podName:b298d9ff-9b10-42f4-aabf-259673933a1a nodeName:}" failed. No retries permitted until 2025-10-10 13:29:50.97667353 +0000 UTC m=+704.874330303 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b298d9ff-9b10-42f4-aabf-259673933a1a-memberlist") pod "speaker-92z9w" (UID: "b298d9ff-9b10-42f4-aabf-259673933a1a") : secret "metallb-memberlist" not found Oct 10 13:29:49 crc kubenswrapper[4745]: I1010 13:29:49.983090 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b298d9ff-9b10-42f4-aabf-259673933a1a-metrics-certs\") pod \"speaker-92z9w\" (UID: \"b298d9ff-9b10-42f4-aabf-259673933a1a\") " pod="metallb-system/speaker-92z9w" Oct 10 13:29:50 crc kubenswrapper[4745]: I1010 13:29:50.121852 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8ktnf" Oct 10 13:29:50 crc kubenswrapper[4745]: I1010 13:29:50.136221 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-52wk2" Oct 10 13:29:50 crc kubenswrapper[4745]: I1010 13:29:50.554905 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-8ktnf"] Oct 10 13:29:50 crc kubenswrapper[4745]: W1010 13:29:50.559266 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda362b7b7_7534_4a4f_ae57_aac84fd292c7.slice/crio-e461943b6625009040ea711a3be8d2d9b734768b533426ea8d615220046e35c2 WatchSource:0}: Error finding container e461943b6625009040ea711a3be8d2d9b734768b533426ea8d615220046e35c2: Status 404 returned error can't find the container with id e461943b6625009040ea711a3be8d2d9b734768b533426ea8d615220046e35c2 Oct 10 13:29:50 crc kubenswrapper[4745]: I1010 13:29:50.935299 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-52wk2" event={"ID":"e3feebb0-e7f8-4aef-a725-2b9cf003364c","Type":"ContainerStarted","Data":"f54602861933b5664b97ecb9f42079ef912517df24ae45e0c3a72cd39493dded"} Oct 10 13:29:50 crc kubenswrapper[4745]: I1010 13:29:50.937273 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-ktgdl" event={"ID":"c1711da0-4710-4569-bdbd-2afe523b91be","Type":"ContainerStarted","Data":"c2d3e55ea976e3f9500213b52b1eff732ceafd61e40029fbf7380d3d9d6d75ad"} Oct 10 13:29:50 crc kubenswrapper[4745]: I1010 13:29:50.937337 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-ktgdl" event={"ID":"c1711da0-4710-4569-bdbd-2afe523b91be","Type":"ContainerStarted","Data":"e7d7a8758022bc27bcc01ab1148e52dfaee82cbeced0422ff87111652df98c62"} Oct 10 13:29:50 crc kubenswrapper[4745]: I1010 13:29:50.937677 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-ktgdl" Oct 10 13:29:50 crc kubenswrapper[4745]: I1010 13:29:50.939009 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8ktnf" event={"ID":"a362b7b7-7534-4a4f-ae57-aac84fd292c7","Type":"ContainerStarted","Data":"e461943b6625009040ea711a3be8d2d9b734768b533426ea8d615220046e35c2"} Oct 10 13:29:50 crc kubenswrapper[4745]: I1010 13:29:50.978369 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-ktgdl" podStartSLOduration=1.9783385409999998 podStartE2EDuration="1.978338541s" podCreationTimestamp="2025-10-10 13:29:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:29:50.968228586 +0000 UTC m=+704.865885379" watchObservedRunningTime="2025-10-10 13:29:50.978338541 +0000 UTC m=+704.875995334" Oct 10 13:29:50 crc kubenswrapper[4745]: I1010 13:29:50.986237 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b298d9ff-9b10-42f4-aabf-259673933a1a-memberlist\") pod \"speaker-92z9w\" (UID: \"b298d9ff-9b10-42f4-aabf-259673933a1a\") " pod="metallb-system/speaker-92z9w" Oct 10 13:29:50 crc kubenswrapper[4745]: I1010 13:29:50.992387 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b298d9ff-9b10-42f4-aabf-259673933a1a-memberlist\") pod \"speaker-92z9w\" (UID: \"b298d9ff-9b10-42f4-aabf-259673933a1a\") " pod="metallb-system/speaker-92z9w" Oct 10 13:29:51 crc kubenswrapper[4745]: I1010 13:29:51.120921 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-92z9w" Oct 10 13:29:51 crc kubenswrapper[4745]: W1010 13:29:51.140279 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb298d9ff_9b10_42f4_aabf_259673933a1a.slice/crio-238a9f5e2b2d36b2b1997bbc504d7818a0beb486c2c49101094f40eaceb17874 WatchSource:0}: Error finding container 238a9f5e2b2d36b2b1997bbc504d7818a0beb486c2c49101094f40eaceb17874: Status 404 returned error can't find the container with id 238a9f5e2b2d36b2b1997bbc504d7818a0beb486c2c49101094f40eaceb17874 Oct 10 13:29:51 crc kubenswrapper[4745]: I1010 13:29:51.949151 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-92z9w" event={"ID":"b298d9ff-9b10-42f4-aabf-259673933a1a","Type":"ContainerStarted","Data":"a73d55985319b8f5f08fc8c55c61ac9fed2c0814ac11d8bd0febde31b4d20c83"} Oct 10 13:29:51 crc kubenswrapper[4745]: I1010 13:29:51.950847 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-92z9w" event={"ID":"b298d9ff-9b10-42f4-aabf-259673933a1a","Type":"ContainerStarted","Data":"c735171b79334805ce54c1089ddd3370b360be38e8942e12b8d7e4243f3b416d"} Oct 10 13:29:51 crc kubenswrapper[4745]: I1010 13:29:51.950893 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-92z9w" event={"ID":"b298d9ff-9b10-42f4-aabf-259673933a1a","Type":"ContainerStarted","Data":"238a9f5e2b2d36b2b1997bbc504d7818a0beb486c2c49101094f40eaceb17874"} Oct 10 13:29:51 crc kubenswrapper[4745]: I1010 13:29:51.951702 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-92z9w" Oct 10 13:29:51 crc kubenswrapper[4745]: I1010 13:29:51.976871 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-92z9w" podStartSLOduration=2.976850608 podStartE2EDuration="2.976850608s" podCreationTimestamp="2025-10-10 13:29:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:29:51.973843749 +0000 UTC m=+705.871500532" watchObservedRunningTime="2025-10-10 13:29:51.976850608 +0000 UTC m=+705.874507381" Oct 10 13:29:57 crc kubenswrapper[4745]: I1010 13:29:57.983128 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8ktnf" event={"ID":"a362b7b7-7534-4a4f-ae57-aac84fd292c7","Type":"ContainerStarted","Data":"60f7a6da09b86a559685070a0a3a3869efb43640b1f58a9dc7df652b97e69019"} Oct 10 13:29:57 crc kubenswrapper[4745]: I1010 13:29:57.984031 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8ktnf" Oct 10 13:29:57 crc kubenswrapper[4745]: I1010 13:29:57.984833 4745 generic.go:334] "Generic (PLEG): container finished" podID="e3feebb0-e7f8-4aef-a725-2b9cf003364c" containerID="c0fa427201d790ac306b5df84b1b7ed1d1c5e58d540b82a291629d112e48a42c" exitCode=0 Oct 10 13:29:57 crc kubenswrapper[4745]: I1010 13:29:57.984860 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-52wk2" event={"ID":"e3feebb0-e7f8-4aef-a725-2b9cf003364c","Type":"ContainerDied","Data":"c0fa427201d790ac306b5df84b1b7ed1d1c5e58d540b82a291629d112e48a42c"} Oct 10 13:29:58 crc kubenswrapper[4745]: I1010 13:29:58.001468 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8ktnf" podStartSLOduration=1.797334589 podStartE2EDuration="9.001446178s" podCreationTimestamp="2025-10-10 13:29:49 +0000 UTC" firstStartedPulling="2025-10-10 13:29:50.560863436 +0000 UTC m=+704.458520199" lastFinishedPulling="2025-10-10 13:29:57.764975025 +0000 UTC m=+711.662631788" observedRunningTime="2025-10-10 13:29:57.996304439 +0000 UTC m=+711.893961212" watchObservedRunningTime="2025-10-10 13:29:58.001446178 +0000 UTC m=+711.899102941" Oct 10 13:29:58 crc kubenswrapper[4745]: I1010 13:29:58.995087 4745 generic.go:334] "Generic (PLEG): container finished" podID="e3feebb0-e7f8-4aef-a725-2b9cf003364c" containerID="dded56a4258f088e7fabdabdb848d6ea8aae678225d85c6dba8fee47a8624f15" exitCode=0 Oct 10 13:29:58 crc kubenswrapper[4745]: I1010 13:29:58.995215 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-52wk2" event={"ID":"e3feebb0-e7f8-4aef-a725-2b9cf003364c","Type":"ContainerDied","Data":"dded56a4258f088e7fabdabdb848d6ea8aae678225d85c6dba8fee47a8624f15"} Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.005933 4745 generic.go:334] "Generic (PLEG): container finished" podID="e3feebb0-e7f8-4aef-a725-2b9cf003364c" containerID="056c401f8ec2895ee77fc63d516f85410cf7fc2ccb13db5371f125abd2f5e905" exitCode=0 Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.005997 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-52wk2" event={"ID":"e3feebb0-e7f8-4aef-a725-2b9cf003364c","Type":"ContainerDied","Data":"056c401f8ec2895ee77fc63d516f85410cf7fc2ccb13db5371f125abd2f5e905"} Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.144474 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5"] Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.146379 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.148629 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.149366 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.156313 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5"] Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.245335 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9eed511e-b79b-47ee-88ec-4fa9fb65146e-config-volume\") pod \"collect-profiles-29335050-ds4l5\" (UID: \"9eed511e-b79b-47ee-88ec-4fa9fb65146e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.245405 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9eed511e-b79b-47ee-88ec-4fa9fb65146e-secret-volume\") pod \"collect-profiles-29335050-ds4l5\" (UID: \"9eed511e-b79b-47ee-88ec-4fa9fb65146e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.245591 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bk9wh\" (UniqueName: \"kubernetes.io/projected/9eed511e-b79b-47ee-88ec-4fa9fb65146e-kube-api-access-bk9wh\") pod \"collect-profiles-29335050-ds4l5\" (UID: \"9eed511e-b79b-47ee-88ec-4fa9fb65146e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.346339 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9eed511e-b79b-47ee-88ec-4fa9fb65146e-config-volume\") pod \"collect-profiles-29335050-ds4l5\" (UID: \"9eed511e-b79b-47ee-88ec-4fa9fb65146e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.346384 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9eed511e-b79b-47ee-88ec-4fa9fb65146e-secret-volume\") pod \"collect-profiles-29335050-ds4l5\" (UID: \"9eed511e-b79b-47ee-88ec-4fa9fb65146e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.346424 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bk9wh\" (UniqueName: \"kubernetes.io/projected/9eed511e-b79b-47ee-88ec-4fa9fb65146e-kube-api-access-bk9wh\") pod \"collect-profiles-29335050-ds4l5\" (UID: \"9eed511e-b79b-47ee-88ec-4fa9fb65146e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.347918 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9eed511e-b79b-47ee-88ec-4fa9fb65146e-config-volume\") pod \"collect-profiles-29335050-ds4l5\" (UID: \"9eed511e-b79b-47ee-88ec-4fa9fb65146e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.353286 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9eed511e-b79b-47ee-88ec-4fa9fb65146e-secret-volume\") pod \"collect-profiles-29335050-ds4l5\" (UID: \"9eed511e-b79b-47ee-88ec-4fa9fb65146e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.365782 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bk9wh\" (UniqueName: \"kubernetes.io/projected/9eed511e-b79b-47ee-88ec-4fa9fb65146e-kube-api-access-bk9wh\") pod \"collect-profiles-29335050-ds4l5\" (UID: \"9eed511e-b79b-47ee-88ec-4fa9fb65146e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.520125 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" Oct 10 13:30:00 crc kubenswrapper[4745]: I1010 13:30:00.755688 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5"] Oct 10 13:30:00 crc kubenswrapper[4745]: W1010 13:30:00.771973 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9eed511e_b79b_47ee_88ec_4fa9fb65146e.slice/crio-fbb5edb109b99a2d5938513e6707e57a0f44d5ac2095a6e76f1fa576faac0bb0 WatchSource:0}: Error finding container fbb5edb109b99a2d5938513e6707e57a0f44d5ac2095a6e76f1fa576faac0bb0: Status 404 returned error can't find the container with id fbb5edb109b99a2d5938513e6707e57a0f44d5ac2095a6e76f1fa576faac0bb0 Oct 10 13:30:01 crc kubenswrapper[4745]: I1010 13:30:01.017091 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-52wk2" event={"ID":"e3feebb0-e7f8-4aef-a725-2b9cf003364c","Type":"ContainerStarted","Data":"ac0c792dc875448675ce3f2b1d612d3285d59c95e5c48ecc0d351d12ab931096"} Oct 10 13:30:01 crc kubenswrapper[4745]: I1010 13:30:01.017127 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-52wk2" event={"ID":"e3feebb0-e7f8-4aef-a725-2b9cf003364c","Type":"ContainerStarted","Data":"5c71eeaf5f6c35773f8439cc822347b6946deb28acc022cad0ffd5c8483b5c22"} Oct 10 13:30:01 crc kubenswrapper[4745]: I1010 13:30:01.017138 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-52wk2" event={"ID":"e3feebb0-e7f8-4aef-a725-2b9cf003364c","Type":"ContainerStarted","Data":"c8700d705b587247973d6be173d57ac19e1bb693aa3f2d40937a3aab4daf187e"} Oct 10 13:30:01 crc kubenswrapper[4745]: I1010 13:30:01.017147 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-52wk2" event={"ID":"e3feebb0-e7f8-4aef-a725-2b9cf003364c","Type":"ContainerStarted","Data":"6ed0af5e7afbef9def3a69501520d2eac78a67d497583c399e0eca855ed3cc5f"} Oct 10 13:30:01 crc kubenswrapper[4745]: I1010 13:30:01.017155 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-52wk2" event={"ID":"e3feebb0-e7f8-4aef-a725-2b9cf003364c","Type":"ContainerStarted","Data":"7ae200d751b85a2941d04e4597a98f8e364525051daecb415dc661273c58845a"} Oct 10 13:30:01 crc kubenswrapper[4745]: I1010 13:30:01.018072 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" event={"ID":"9eed511e-b79b-47ee-88ec-4fa9fb65146e","Type":"ContainerStarted","Data":"293b6ca4af800b62d70c93b79a536134a5f6b42a1aad6c0db460acfdd3e51754"} Oct 10 13:30:01 crc kubenswrapper[4745]: I1010 13:30:01.018091 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" event={"ID":"9eed511e-b79b-47ee-88ec-4fa9fb65146e","Type":"ContainerStarted","Data":"fbb5edb109b99a2d5938513e6707e57a0f44d5ac2095a6e76f1fa576faac0bb0"} Oct 10 13:30:01 crc kubenswrapper[4745]: I1010 13:30:01.039508 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" podStartSLOduration=1.03949206 podStartE2EDuration="1.03949206s" podCreationTimestamp="2025-10-10 13:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:30:01.036197734 +0000 UTC m=+714.933854507" watchObservedRunningTime="2025-10-10 13:30:01.03949206 +0000 UTC m=+714.937148823" Oct 10 13:30:01 crc kubenswrapper[4745]: I1010 13:30:01.124997 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-92z9w" Oct 10 13:30:02 crc kubenswrapper[4745]: I1010 13:30:02.027435 4745 generic.go:334] "Generic (PLEG): container finished" podID="9eed511e-b79b-47ee-88ec-4fa9fb65146e" containerID="293b6ca4af800b62d70c93b79a536134a5f6b42a1aad6c0db460acfdd3e51754" exitCode=0 Oct 10 13:30:02 crc kubenswrapper[4745]: I1010 13:30:02.027488 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" event={"ID":"9eed511e-b79b-47ee-88ec-4fa9fb65146e","Type":"ContainerDied","Data":"293b6ca4af800b62d70c93b79a536134a5f6b42a1aad6c0db460acfdd3e51754"} Oct 10 13:30:02 crc kubenswrapper[4745]: I1010 13:30:02.033745 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-52wk2" event={"ID":"e3feebb0-e7f8-4aef-a725-2b9cf003364c","Type":"ContainerStarted","Data":"fa8ff1baf358c6e2f048bf0c80b93b2d74dd1ec7aa28d63eeca5bcf5ceacedf7"} Oct 10 13:30:02 crc kubenswrapper[4745]: I1010 13:30:02.034045 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-52wk2" Oct 10 13:30:02 crc kubenswrapper[4745]: I1010 13:30:02.074106 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-52wk2" podStartSLOduration=5.597207835 podStartE2EDuration="13.074091923s" podCreationTimestamp="2025-10-10 13:29:49 +0000 UTC" firstStartedPulling="2025-10-10 13:29:50.265879781 +0000 UTC m=+704.163536544" lastFinishedPulling="2025-10-10 13:29:57.742763869 +0000 UTC m=+711.640420632" observedRunningTime="2025-10-10 13:30:02.073899388 +0000 UTC m=+715.971556171" watchObservedRunningTime="2025-10-10 13:30:02.074091923 +0000 UTC m=+715.971748686" Oct 10 13:30:03 crc kubenswrapper[4745]: I1010 13:30:03.362529 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" Oct 10 13:30:03 crc kubenswrapper[4745]: I1010 13:30:03.488150 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9eed511e-b79b-47ee-88ec-4fa9fb65146e-config-volume\") pod \"9eed511e-b79b-47ee-88ec-4fa9fb65146e\" (UID: \"9eed511e-b79b-47ee-88ec-4fa9fb65146e\") " Oct 10 13:30:03 crc kubenswrapper[4745]: I1010 13:30:03.488233 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bk9wh\" (UniqueName: \"kubernetes.io/projected/9eed511e-b79b-47ee-88ec-4fa9fb65146e-kube-api-access-bk9wh\") pod \"9eed511e-b79b-47ee-88ec-4fa9fb65146e\" (UID: \"9eed511e-b79b-47ee-88ec-4fa9fb65146e\") " Oct 10 13:30:03 crc kubenswrapper[4745]: I1010 13:30:03.488305 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9eed511e-b79b-47ee-88ec-4fa9fb65146e-secret-volume\") pod \"9eed511e-b79b-47ee-88ec-4fa9fb65146e\" (UID: \"9eed511e-b79b-47ee-88ec-4fa9fb65146e\") " Oct 10 13:30:03 crc kubenswrapper[4745]: I1010 13:30:03.488810 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9eed511e-b79b-47ee-88ec-4fa9fb65146e-config-volume" (OuterVolumeSpecName: "config-volume") pod "9eed511e-b79b-47ee-88ec-4fa9fb65146e" (UID: "9eed511e-b79b-47ee-88ec-4fa9fb65146e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:30:03 crc kubenswrapper[4745]: I1010 13:30:03.494136 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9eed511e-b79b-47ee-88ec-4fa9fb65146e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9eed511e-b79b-47ee-88ec-4fa9fb65146e" (UID: "9eed511e-b79b-47ee-88ec-4fa9fb65146e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:30:03 crc kubenswrapper[4745]: I1010 13:30:03.496129 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eed511e-b79b-47ee-88ec-4fa9fb65146e-kube-api-access-bk9wh" (OuterVolumeSpecName: "kube-api-access-bk9wh") pod "9eed511e-b79b-47ee-88ec-4fa9fb65146e" (UID: "9eed511e-b79b-47ee-88ec-4fa9fb65146e"). InnerVolumeSpecName "kube-api-access-bk9wh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:30:03 crc kubenswrapper[4745]: I1010 13:30:03.589533 4745 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9eed511e-b79b-47ee-88ec-4fa9fb65146e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:03 crc kubenswrapper[4745]: I1010 13:30:03.589565 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bk9wh\" (UniqueName: \"kubernetes.io/projected/9eed511e-b79b-47ee-88ec-4fa9fb65146e-kube-api-access-bk9wh\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:03 crc kubenswrapper[4745]: I1010 13:30:03.589576 4745 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9eed511e-b79b-47ee-88ec-4fa9fb65146e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:04 crc kubenswrapper[4745]: I1010 13:30:04.061690 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" event={"ID":"9eed511e-b79b-47ee-88ec-4fa9fb65146e","Type":"ContainerDied","Data":"fbb5edb109b99a2d5938513e6707e57a0f44d5ac2095a6e76f1fa576faac0bb0"} Oct 10 13:30:04 crc kubenswrapper[4745]: I1010 13:30:04.061767 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbb5edb109b99a2d5938513e6707e57a0f44d5ac2095a6e76f1fa576faac0bb0" Oct 10 13:30:04 crc kubenswrapper[4745]: I1010 13:30:04.061846 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5" Oct 10 13:30:04 crc kubenswrapper[4745]: I1010 13:30:04.156576 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-6tszb"] Oct 10 13:30:04 crc kubenswrapper[4745]: E1010 13:30:04.156857 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eed511e-b79b-47ee-88ec-4fa9fb65146e" containerName="collect-profiles" Oct 10 13:30:04 crc kubenswrapper[4745]: I1010 13:30:04.156873 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eed511e-b79b-47ee-88ec-4fa9fb65146e" containerName="collect-profiles" Oct 10 13:30:04 crc kubenswrapper[4745]: I1010 13:30:04.157003 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eed511e-b79b-47ee-88ec-4fa9fb65146e" containerName="collect-profiles" Oct 10 13:30:04 crc kubenswrapper[4745]: I1010 13:30:04.157413 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6tszb" Oct 10 13:30:04 crc kubenswrapper[4745]: I1010 13:30:04.160311 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 10 13:30:04 crc kubenswrapper[4745]: I1010 13:30:04.160476 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-c5bwf" Oct 10 13:30:04 crc kubenswrapper[4745]: I1010 13:30:04.160614 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 10 13:30:04 crc kubenswrapper[4745]: I1010 13:30:04.171080 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6tszb"] Oct 10 13:30:04 crc kubenswrapper[4745]: I1010 13:30:04.212544 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4cxg\" (UniqueName: \"kubernetes.io/projected/40d5edac-be93-424b-8f39-1609f8b7cef4-kube-api-access-j4cxg\") pod \"openstack-operator-index-6tszb\" (UID: \"40d5edac-be93-424b-8f39-1609f8b7cef4\") " pod="openstack-operators/openstack-operator-index-6tszb" Oct 10 13:30:04 crc kubenswrapper[4745]: I1010 13:30:04.313885 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4cxg\" (UniqueName: \"kubernetes.io/projected/40d5edac-be93-424b-8f39-1609f8b7cef4-kube-api-access-j4cxg\") pod \"openstack-operator-index-6tszb\" (UID: \"40d5edac-be93-424b-8f39-1609f8b7cef4\") " pod="openstack-operators/openstack-operator-index-6tszb" Oct 10 13:30:04 crc kubenswrapper[4745]: I1010 13:30:04.329417 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4cxg\" (UniqueName: \"kubernetes.io/projected/40d5edac-be93-424b-8f39-1609f8b7cef4-kube-api-access-j4cxg\") pod \"openstack-operator-index-6tszb\" (UID: \"40d5edac-be93-424b-8f39-1609f8b7cef4\") " pod="openstack-operators/openstack-operator-index-6tszb" Oct 10 13:30:04 crc kubenswrapper[4745]: I1010 13:30:04.472290 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6tszb" Oct 10 13:30:04 crc kubenswrapper[4745]: I1010 13:30:04.719504 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6tszb"] Oct 10 13:30:04 crc kubenswrapper[4745]: W1010 13:30:04.727638 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40d5edac_be93_424b_8f39_1609f8b7cef4.slice/crio-0661ff5974074658c21cfc8b335541f0e9018481799909aa842362f1a83b31dd WatchSource:0}: Error finding container 0661ff5974074658c21cfc8b335541f0e9018481799909aa842362f1a83b31dd: Status 404 returned error can't find the container with id 0661ff5974074658c21cfc8b335541f0e9018481799909aa842362f1a83b31dd Oct 10 13:30:05 crc kubenswrapper[4745]: I1010 13:30:05.069087 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6tszb" event={"ID":"40d5edac-be93-424b-8f39-1609f8b7cef4","Type":"ContainerStarted","Data":"0661ff5974074658c21cfc8b335541f0e9018481799909aa842362f1a83b31dd"} Oct 10 13:30:05 crc kubenswrapper[4745]: I1010 13:30:05.136905 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-52wk2" Oct 10 13:30:05 crc kubenswrapper[4745]: I1010 13:30:05.184712 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-52wk2" Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.093258 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6tszb" event={"ID":"40d5edac-be93-424b-8f39-1609f8b7cef4","Type":"ContainerStarted","Data":"5c84a52c7982accfa0d77b4ad57aa44f8865844b951e3820839c54effeab07d1"} Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.126271 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-6tszb" podStartSLOduration=1.483218623 podStartE2EDuration="4.126238441s" podCreationTimestamp="2025-10-10 13:30:04 +0000 UTC" firstStartedPulling="2025-10-10 13:30:04.731558761 +0000 UTC m=+718.629215534" lastFinishedPulling="2025-10-10 13:30:07.374578589 +0000 UTC m=+721.272235352" observedRunningTime="2025-10-10 13:30:08.11444152 +0000 UTC m=+722.012098353" watchObservedRunningTime="2025-10-10 13:30:08.126238441 +0000 UTC m=+722.023895254" Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.423825 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9q9xb"] Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.424070 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" podUID="0088430d-3a96-46ea-a149-7c59eb24d706" containerName="controller-manager" containerID="cri-o://053f3395c80cb98947103d8af9aa61e1ced574f401c42d7560e7d02647094611" gracePeriod=30 Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.549603 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl"] Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.549980 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" podUID="f9365b5c-a2b4-464c-aef6-afd71c5d3427" containerName="route-controller-manager" containerID="cri-o://0a9cbdd2ccc2e6375518e7ba8ed3010b81b21e41e1dcd3e3d51391b2a370a93b" gracePeriod=30 Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.795057 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.868681 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-client-ca\") pod \"0088430d-3a96-46ea-a149-7c59eb24d706\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.868752 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0088430d-3a96-46ea-a149-7c59eb24d706-serving-cert\") pod \"0088430d-3a96-46ea-a149-7c59eb24d706\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.868785 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqbfm\" (UniqueName: \"kubernetes.io/projected/0088430d-3a96-46ea-a149-7c59eb24d706-kube-api-access-gqbfm\") pod \"0088430d-3a96-46ea-a149-7c59eb24d706\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.868814 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-proxy-ca-bundles\") pod \"0088430d-3a96-46ea-a149-7c59eb24d706\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.868878 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-config\") pod \"0088430d-3a96-46ea-a149-7c59eb24d706\" (UID: \"0088430d-3a96-46ea-a149-7c59eb24d706\") " Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.869697 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "0088430d-3a96-46ea-a149-7c59eb24d706" (UID: "0088430d-3a96-46ea-a149-7c59eb24d706"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.869741 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-client-ca" (OuterVolumeSpecName: "client-ca") pod "0088430d-3a96-46ea-a149-7c59eb24d706" (UID: "0088430d-3a96-46ea-a149-7c59eb24d706"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.869750 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-config" (OuterVolumeSpecName: "config") pod "0088430d-3a96-46ea-a149-7c59eb24d706" (UID: "0088430d-3a96-46ea-a149-7c59eb24d706"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.875626 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0088430d-3a96-46ea-a149-7c59eb24d706-kube-api-access-gqbfm" (OuterVolumeSpecName: "kube-api-access-gqbfm") pod "0088430d-3a96-46ea-a149-7c59eb24d706" (UID: "0088430d-3a96-46ea-a149-7c59eb24d706"). InnerVolumeSpecName "kube-api-access-gqbfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.876210 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0088430d-3a96-46ea-a149-7c59eb24d706-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0088430d-3a96-46ea-a149-7c59eb24d706" (UID: "0088430d-3a96-46ea-a149-7c59eb24d706"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.879307 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.970125 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9365b5c-a2b4-464c-aef6-afd71c5d3427-client-ca\") pod \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\" (UID: \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\") " Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.970230 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9365b5c-a2b4-464c-aef6-afd71c5d3427-config\") pod \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\" (UID: \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\") " Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.970323 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zgdj\" (UniqueName: \"kubernetes.io/projected/f9365b5c-a2b4-464c-aef6-afd71c5d3427-kube-api-access-5zgdj\") pod \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\" (UID: \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\") " Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.970350 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9365b5c-a2b4-464c-aef6-afd71c5d3427-serving-cert\") pod \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\" (UID: \"f9365b5c-a2b4-464c-aef6-afd71c5d3427\") " Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.970671 4745 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-client-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.970689 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0088430d-3a96-46ea-a149-7c59eb24d706-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.970700 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqbfm\" (UniqueName: \"kubernetes.io/projected/0088430d-3a96-46ea-a149-7c59eb24d706-kube-api-access-gqbfm\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.970712 4745 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.970725 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0088430d-3a96-46ea-a149-7c59eb24d706-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.970865 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9365b5c-a2b4-464c-aef6-afd71c5d3427-client-ca" (OuterVolumeSpecName: "client-ca") pod "f9365b5c-a2b4-464c-aef6-afd71c5d3427" (UID: "f9365b5c-a2b4-464c-aef6-afd71c5d3427"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.971024 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9365b5c-a2b4-464c-aef6-afd71c5d3427-config" (OuterVolumeSpecName: "config") pod "f9365b5c-a2b4-464c-aef6-afd71c5d3427" (UID: "f9365b5c-a2b4-464c-aef6-afd71c5d3427"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.973620 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9365b5c-a2b4-464c-aef6-afd71c5d3427-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f9365b5c-a2b4-464c-aef6-afd71c5d3427" (UID: "f9365b5c-a2b4-464c-aef6-afd71c5d3427"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:30:08 crc kubenswrapper[4745]: I1010 13:30:08.973671 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9365b5c-a2b4-464c-aef6-afd71c5d3427-kube-api-access-5zgdj" (OuterVolumeSpecName: "kube-api-access-5zgdj") pod "f9365b5c-a2b4-464c-aef6-afd71c5d3427" (UID: "f9365b5c-a2b4-464c-aef6-afd71c5d3427"). InnerVolumeSpecName "kube-api-access-5zgdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.071510 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zgdj\" (UniqueName: \"kubernetes.io/projected/f9365b5c-a2b4-464c-aef6-afd71c5d3427-kube-api-access-5zgdj\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.071549 4745 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9365b5c-a2b4-464c-aef6-afd71c5d3427-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.071563 4745 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9365b5c-a2b4-464c-aef6-afd71c5d3427-client-ca\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.071574 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9365b5c-a2b4-464c-aef6-afd71c5d3427-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.099204 4745 generic.go:334] "Generic (PLEG): container finished" podID="f9365b5c-a2b4-464c-aef6-afd71c5d3427" containerID="0a9cbdd2ccc2e6375518e7ba8ed3010b81b21e41e1dcd3e3d51391b2a370a93b" exitCode=0 Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.099238 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.099307 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" event={"ID":"f9365b5c-a2b4-464c-aef6-afd71c5d3427","Type":"ContainerDied","Data":"0a9cbdd2ccc2e6375518e7ba8ed3010b81b21e41e1dcd3e3d51391b2a370a93b"} Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.099371 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl" event={"ID":"f9365b5c-a2b4-464c-aef6-afd71c5d3427","Type":"ContainerDied","Data":"6b61c5ef92d84df5681d150821515793325c11f124c3b6fd0d9b7cab2585f349"} Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.099395 4745 scope.go:117] "RemoveContainer" containerID="0a9cbdd2ccc2e6375518e7ba8ed3010b81b21e41e1dcd3e3d51391b2a370a93b" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.102349 4745 generic.go:334] "Generic (PLEG): container finished" podID="0088430d-3a96-46ea-a149-7c59eb24d706" containerID="053f3395c80cb98947103d8af9aa61e1ced574f401c42d7560e7d02647094611" exitCode=0 Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.102387 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.102428 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" event={"ID":"0088430d-3a96-46ea-a149-7c59eb24d706","Type":"ContainerDied","Data":"053f3395c80cb98947103d8af9aa61e1ced574f401c42d7560e7d02647094611"} Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.102465 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9q9xb" event={"ID":"0088430d-3a96-46ea-a149-7c59eb24d706","Type":"ContainerDied","Data":"0242f4f578af6c17968a8008de59ed305086c659e8032c1930a698c69e1bf77c"} Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.115296 4745 scope.go:117] "RemoveContainer" containerID="0a9cbdd2ccc2e6375518e7ba8ed3010b81b21e41e1dcd3e3d51391b2a370a93b" Oct 10 13:30:09 crc kubenswrapper[4745]: E1010 13:30:09.115709 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a9cbdd2ccc2e6375518e7ba8ed3010b81b21e41e1dcd3e3d51391b2a370a93b\": container with ID starting with 0a9cbdd2ccc2e6375518e7ba8ed3010b81b21e41e1dcd3e3d51391b2a370a93b not found: ID does not exist" containerID="0a9cbdd2ccc2e6375518e7ba8ed3010b81b21e41e1dcd3e3d51391b2a370a93b" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.115822 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a9cbdd2ccc2e6375518e7ba8ed3010b81b21e41e1dcd3e3d51391b2a370a93b"} err="failed to get container status \"0a9cbdd2ccc2e6375518e7ba8ed3010b81b21e41e1dcd3e3d51391b2a370a93b\": rpc error: code = NotFound desc = could not find container \"0a9cbdd2ccc2e6375518e7ba8ed3010b81b21e41e1dcd3e3d51391b2a370a93b\": container with ID starting with 0a9cbdd2ccc2e6375518e7ba8ed3010b81b21e41e1dcd3e3d51391b2a370a93b not found: ID does not exist" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.115876 4745 scope.go:117] "RemoveContainer" containerID="053f3395c80cb98947103d8af9aa61e1ced574f401c42d7560e7d02647094611" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.134570 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl"] Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.141433 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nwvhl"] Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.148919 4745 scope.go:117] "RemoveContainer" containerID="053f3395c80cb98947103d8af9aa61e1ced574f401c42d7560e7d02647094611" Oct 10 13:30:09 crc kubenswrapper[4745]: E1010 13:30:09.149371 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"053f3395c80cb98947103d8af9aa61e1ced574f401c42d7560e7d02647094611\": container with ID starting with 053f3395c80cb98947103d8af9aa61e1ced574f401c42d7560e7d02647094611 not found: ID does not exist" containerID="053f3395c80cb98947103d8af9aa61e1ced574f401c42d7560e7d02647094611" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.149411 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"053f3395c80cb98947103d8af9aa61e1ced574f401c42d7560e7d02647094611"} err="failed to get container status \"053f3395c80cb98947103d8af9aa61e1ced574f401c42d7560e7d02647094611\": rpc error: code = NotFound desc = could not find container \"053f3395c80cb98947103d8af9aa61e1ced574f401c42d7560e7d02647094611\": container with ID starting with 053f3395c80cb98947103d8af9aa61e1ced574f401c42d7560e7d02647094611 not found: ID does not exist" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.150133 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9q9xb"] Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.154351 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9q9xb"] Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.636779 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-ktgdl" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.818189 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-68597564d4-4blcm"] Oct 10 13:30:09 crc kubenswrapper[4745]: E1010 13:30:09.818436 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9365b5c-a2b4-464c-aef6-afd71c5d3427" containerName="route-controller-manager" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.818452 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9365b5c-a2b4-464c-aef6-afd71c5d3427" containerName="route-controller-manager" Oct 10 13:30:09 crc kubenswrapper[4745]: E1010 13:30:09.818484 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0088430d-3a96-46ea-a149-7c59eb24d706" containerName="controller-manager" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.818491 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="0088430d-3a96-46ea-a149-7c59eb24d706" containerName="controller-manager" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.818610 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9365b5c-a2b4-464c-aef6-afd71c5d3427" containerName="route-controller-manager" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.818626 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="0088430d-3a96-46ea-a149-7c59eb24d706" containerName="controller-manager" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.819067 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.823877 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.824189 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.824979 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.825050 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.825418 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.825479 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.827823 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95"] Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.828549 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.830716 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.834852 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.838168 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.838582 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.839454 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.841676 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-68597564d4-4blcm"] Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.857118 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95"] Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.862478 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.885930 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.888581 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcr62\" (UniqueName: \"kubernetes.io/projected/2f9f7bca-8353-435a-9293-d6a9459f1306-kube-api-access-kcr62\") pod \"controller-manager-68597564d4-4blcm\" (UID: \"2f9f7bca-8353-435a-9293-d6a9459f1306\") " pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.888644 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2f9f7bca-8353-435a-9293-d6a9459f1306-client-ca\") pod \"controller-manager-68597564d4-4blcm\" (UID: \"2f9f7bca-8353-435a-9293-d6a9459f1306\") " pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.888671 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f9f7bca-8353-435a-9293-d6a9459f1306-serving-cert\") pod \"controller-manager-68597564d4-4blcm\" (UID: \"2f9f7bca-8353-435a-9293-d6a9459f1306\") " pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.888711 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f9f7bca-8353-435a-9293-d6a9459f1306-config\") pod \"controller-manager-68597564d4-4blcm\" (UID: \"2f9f7bca-8353-435a-9293-d6a9459f1306\") " pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.888768 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2f9f7bca-8353-435a-9293-d6a9459f1306-proxy-ca-bundles\") pod \"controller-manager-68597564d4-4blcm\" (UID: \"2f9f7bca-8353-435a-9293-d6a9459f1306\") " pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.989982 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7aea87b5-85a8-486d-bee8-4602b93cc95f-client-ca\") pod \"route-controller-manager-5d586496d9-9mv95\" (UID: \"7aea87b5-85a8-486d-bee8-4602b93cc95f\") " pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.990077 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7aea87b5-85a8-486d-bee8-4602b93cc95f-serving-cert\") pod \"route-controller-manager-5d586496d9-9mv95\" (UID: \"7aea87b5-85a8-486d-bee8-4602b93cc95f\") " pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.990115 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w2g6\" (UniqueName: \"kubernetes.io/projected/7aea87b5-85a8-486d-bee8-4602b93cc95f-kube-api-access-9w2g6\") pod \"route-controller-manager-5d586496d9-9mv95\" (UID: \"7aea87b5-85a8-486d-bee8-4602b93cc95f\") " pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.990150 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aea87b5-85a8-486d-bee8-4602b93cc95f-config\") pod \"route-controller-manager-5d586496d9-9mv95\" (UID: \"7aea87b5-85a8-486d-bee8-4602b93cc95f\") " pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.990247 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcr62\" (UniqueName: \"kubernetes.io/projected/2f9f7bca-8353-435a-9293-d6a9459f1306-kube-api-access-kcr62\") pod \"controller-manager-68597564d4-4blcm\" (UID: \"2f9f7bca-8353-435a-9293-d6a9459f1306\") " pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.990419 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2f9f7bca-8353-435a-9293-d6a9459f1306-client-ca\") pod \"controller-manager-68597564d4-4blcm\" (UID: \"2f9f7bca-8353-435a-9293-d6a9459f1306\") " pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.990445 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f9f7bca-8353-435a-9293-d6a9459f1306-serving-cert\") pod \"controller-manager-68597564d4-4blcm\" (UID: \"2f9f7bca-8353-435a-9293-d6a9459f1306\") " pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.990482 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f9f7bca-8353-435a-9293-d6a9459f1306-config\") pod \"controller-manager-68597564d4-4blcm\" (UID: \"2f9f7bca-8353-435a-9293-d6a9459f1306\") " pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.990515 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2f9f7bca-8353-435a-9293-d6a9459f1306-proxy-ca-bundles\") pod \"controller-manager-68597564d4-4blcm\" (UID: \"2f9f7bca-8353-435a-9293-d6a9459f1306\") " pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.991989 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2f9f7bca-8353-435a-9293-d6a9459f1306-proxy-ca-bundles\") pod \"controller-manager-68597564d4-4blcm\" (UID: \"2f9f7bca-8353-435a-9293-d6a9459f1306\") " pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.992239 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f9f7bca-8353-435a-9293-d6a9459f1306-config\") pod \"controller-manager-68597564d4-4blcm\" (UID: \"2f9f7bca-8353-435a-9293-d6a9459f1306\") " pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.992398 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2f9f7bca-8353-435a-9293-d6a9459f1306-client-ca\") pod \"controller-manager-68597564d4-4blcm\" (UID: \"2f9f7bca-8353-435a-9293-d6a9459f1306\") " pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:09 crc kubenswrapper[4745]: I1010 13:30:09.995641 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f9f7bca-8353-435a-9293-d6a9459f1306-serving-cert\") pod \"controller-manager-68597564d4-4blcm\" (UID: \"2f9f7bca-8353-435a-9293-d6a9459f1306\") " pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:10 crc kubenswrapper[4745]: I1010 13:30:10.011659 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcr62\" (UniqueName: \"kubernetes.io/projected/2f9f7bca-8353-435a-9293-d6a9459f1306-kube-api-access-kcr62\") pod \"controller-manager-68597564d4-4blcm\" (UID: \"2f9f7bca-8353-435a-9293-d6a9459f1306\") " pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:10 crc kubenswrapper[4745]: I1010 13:30:10.091537 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7aea87b5-85a8-486d-bee8-4602b93cc95f-client-ca\") pod \"route-controller-manager-5d586496d9-9mv95\" (UID: \"7aea87b5-85a8-486d-bee8-4602b93cc95f\") " pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" Oct 10 13:30:10 crc kubenswrapper[4745]: I1010 13:30:10.091605 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7aea87b5-85a8-486d-bee8-4602b93cc95f-serving-cert\") pod \"route-controller-manager-5d586496d9-9mv95\" (UID: \"7aea87b5-85a8-486d-bee8-4602b93cc95f\") " pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" Oct 10 13:30:10 crc kubenswrapper[4745]: I1010 13:30:10.091647 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w2g6\" (UniqueName: \"kubernetes.io/projected/7aea87b5-85a8-486d-bee8-4602b93cc95f-kube-api-access-9w2g6\") pod \"route-controller-manager-5d586496d9-9mv95\" (UID: \"7aea87b5-85a8-486d-bee8-4602b93cc95f\") " pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" Oct 10 13:30:10 crc kubenswrapper[4745]: I1010 13:30:10.091746 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aea87b5-85a8-486d-bee8-4602b93cc95f-config\") pod \"route-controller-manager-5d586496d9-9mv95\" (UID: \"7aea87b5-85a8-486d-bee8-4602b93cc95f\") " pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" Oct 10 13:30:10 crc kubenswrapper[4745]: I1010 13:30:10.092998 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7aea87b5-85a8-486d-bee8-4602b93cc95f-client-ca\") pod \"route-controller-manager-5d586496d9-9mv95\" (UID: \"7aea87b5-85a8-486d-bee8-4602b93cc95f\") " pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" Oct 10 13:30:10 crc kubenswrapper[4745]: I1010 13:30:10.093403 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aea87b5-85a8-486d-bee8-4602b93cc95f-config\") pod \"route-controller-manager-5d586496d9-9mv95\" (UID: \"7aea87b5-85a8-486d-bee8-4602b93cc95f\") " pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" Oct 10 13:30:10 crc kubenswrapper[4745]: I1010 13:30:10.104655 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7aea87b5-85a8-486d-bee8-4602b93cc95f-serving-cert\") pod \"route-controller-manager-5d586496d9-9mv95\" (UID: \"7aea87b5-85a8-486d-bee8-4602b93cc95f\") " pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" Oct 10 13:30:10 crc kubenswrapper[4745]: I1010 13:30:10.113528 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w2g6\" (UniqueName: \"kubernetes.io/projected/7aea87b5-85a8-486d-bee8-4602b93cc95f-kube-api-access-9w2g6\") pod \"route-controller-manager-5d586496d9-9mv95\" (UID: \"7aea87b5-85a8-486d-bee8-4602b93cc95f\") " pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" Oct 10 13:30:10 crc kubenswrapper[4745]: I1010 13:30:10.130687 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8ktnf" Oct 10 13:30:10 crc kubenswrapper[4745]: I1010 13:30:10.140012 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-52wk2" Oct 10 13:30:10 crc kubenswrapper[4745]: I1010 13:30:10.192054 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:10 crc kubenswrapper[4745]: I1010 13:30:10.209879 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" Oct 10 13:30:10 crc kubenswrapper[4745]: I1010 13:30:10.712459 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-68597564d4-4blcm"] Oct 10 13:30:10 crc kubenswrapper[4745]: W1010 13:30:10.720077 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f9f7bca_8353_435a_9293_d6a9459f1306.slice/crio-cbab9a0acecd4d813763ddf0484d6c4d60633eea588da8cb2ffbc5b2ac58838e WatchSource:0}: Error finding container cbab9a0acecd4d813763ddf0484d6c4d60633eea588da8cb2ffbc5b2ac58838e: Status 404 returned error can't find the container with id cbab9a0acecd4d813763ddf0484d6c4d60633eea588da8cb2ffbc5b2ac58838e Oct 10 13:30:10 crc kubenswrapper[4745]: I1010 13:30:10.755623 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0088430d-3a96-46ea-a149-7c59eb24d706" path="/var/lib/kubelet/pods/0088430d-3a96-46ea-a149-7c59eb24d706/volumes" Oct 10 13:30:10 crc kubenswrapper[4745]: I1010 13:30:10.756190 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9365b5c-a2b4-464c-aef6-afd71c5d3427" path="/var/lib/kubelet/pods/f9365b5c-a2b4-464c-aef6-afd71c5d3427/volumes" Oct 10 13:30:10 crc kubenswrapper[4745]: I1010 13:30:10.756689 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95"] Oct 10 13:30:10 crc kubenswrapper[4745]: W1010 13:30:10.763360 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7aea87b5_85a8_486d_bee8_4602b93cc95f.slice/crio-12841eda44316aa14f99df08a64982b74a334ff725428a9565a21684c952d2a1 WatchSource:0}: Error finding container 12841eda44316aa14f99df08a64982b74a334ff725428a9565a21684c952d2a1: Status 404 returned error can't find the container with id 12841eda44316aa14f99df08a64982b74a334ff725428a9565a21684c952d2a1 Oct 10 13:30:11 crc kubenswrapper[4745]: I1010 13:30:11.119813 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" event={"ID":"7aea87b5-85a8-486d-bee8-4602b93cc95f","Type":"ContainerStarted","Data":"5b1bdb9dcca26bcdb29930654983139c63a57e06b10a0877730ad7d1d1c8a2e8"} Oct 10 13:30:11 crc kubenswrapper[4745]: I1010 13:30:11.120176 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" Oct 10 13:30:11 crc kubenswrapper[4745]: I1010 13:30:11.120191 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" event={"ID":"7aea87b5-85a8-486d-bee8-4602b93cc95f","Type":"ContainerStarted","Data":"12841eda44316aa14f99df08a64982b74a334ff725428a9565a21684c952d2a1"} Oct 10 13:30:11 crc kubenswrapper[4745]: I1010 13:30:11.121188 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" event={"ID":"2f9f7bca-8353-435a-9293-d6a9459f1306","Type":"ContainerStarted","Data":"4a71e6532264ea9d572ff4366964d06b3283cd908b30cc396b2245f75304a5c1"} Oct 10 13:30:11 crc kubenswrapper[4745]: I1010 13:30:11.121225 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" event={"ID":"2f9f7bca-8353-435a-9293-d6a9459f1306","Type":"ContainerStarted","Data":"cbab9a0acecd4d813763ddf0484d6c4d60633eea588da8cb2ffbc5b2ac58838e"} Oct 10 13:30:11 crc kubenswrapper[4745]: I1010 13:30:11.121388 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:11 crc kubenswrapper[4745]: I1010 13:30:11.122380 4745 patch_prober.go:28] interesting pod/controller-manager-68597564d4-4blcm container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" start-of-body= Oct 10 13:30:11 crc kubenswrapper[4745]: I1010 13:30:11.122420 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" podUID="2f9f7bca-8353-435a-9293-d6a9459f1306" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" Oct 10 13:30:11 crc kubenswrapper[4745]: I1010 13:30:11.137367 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" podStartSLOduration=3.137351592 podStartE2EDuration="3.137351592s" podCreationTimestamp="2025-10-10 13:30:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:30:11.134774083 +0000 UTC m=+725.032430846" watchObservedRunningTime="2025-10-10 13:30:11.137351592 +0000 UTC m=+725.035008355" Oct 10 13:30:11 crc kubenswrapper[4745]: I1010 13:30:11.157035 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" podStartSLOduration=3.157020443 podStartE2EDuration="3.157020443s" podCreationTimestamp="2025-10-10 13:30:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:30:11.153125623 +0000 UTC m=+725.050782386" watchObservedRunningTime="2025-10-10 13:30:11.157020443 +0000 UTC m=+725.054677206" Oct 10 13:30:11 crc kubenswrapper[4745]: I1010 13:30:11.531315 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5d586496d9-9mv95" Oct 10 13:30:12 crc kubenswrapper[4745]: I1010 13:30:12.131574 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-68597564d4-4blcm" Oct 10 13:30:14 crc kubenswrapper[4745]: I1010 13:30:14.473442 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-6tszb" Oct 10 13:30:14 crc kubenswrapper[4745]: I1010 13:30:14.474252 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-6tszb" Oct 10 13:30:14 crc kubenswrapper[4745]: I1010 13:30:14.504353 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-6tszb" Oct 10 13:30:15 crc kubenswrapper[4745]: I1010 13:30:15.176889 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-6tszb" Oct 10 13:30:15 crc kubenswrapper[4745]: I1010 13:30:15.841303 4745 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 10 13:30:16 crc kubenswrapper[4745]: I1010 13:30:16.102503 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc"] Oct 10 13:30:16 crc kubenswrapper[4745]: I1010 13:30:16.103655 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" Oct 10 13:30:16 crc kubenswrapper[4745]: I1010 13:30:16.107171 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-gn4z9" Oct 10 13:30:16 crc kubenswrapper[4745]: I1010 13:30:16.120892 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc"] Oct 10 13:30:16 crc kubenswrapper[4745]: I1010 13:30:16.165818 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/128069b5-6f89-4da5-9e35-bfe14f1ebd50-util\") pod \"73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc\" (UID: \"128069b5-6f89-4da5-9e35-bfe14f1ebd50\") " pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" Oct 10 13:30:16 crc kubenswrapper[4745]: I1010 13:30:16.166037 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b797b\" (UniqueName: \"kubernetes.io/projected/128069b5-6f89-4da5-9e35-bfe14f1ebd50-kube-api-access-b797b\") pod \"73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc\" (UID: \"128069b5-6f89-4da5-9e35-bfe14f1ebd50\") " pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" Oct 10 13:30:16 crc kubenswrapper[4745]: I1010 13:30:16.166075 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/128069b5-6f89-4da5-9e35-bfe14f1ebd50-bundle\") pod \"73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc\" (UID: \"128069b5-6f89-4da5-9e35-bfe14f1ebd50\") " pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" Oct 10 13:30:16 crc kubenswrapper[4745]: I1010 13:30:16.187321 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:30:16 crc kubenswrapper[4745]: I1010 13:30:16.187368 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:30:16 crc kubenswrapper[4745]: I1010 13:30:16.267198 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b797b\" (UniqueName: \"kubernetes.io/projected/128069b5-6f89-4da5-9e35-bfe14f1ebd50-kube-api-access-b797b\") pod \"73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc\" (UID: \"128069b5-6f89-4da5-9e35-bfe14f1ebd50\") " pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" Oct 10 13:30:16 crc kubenswrapper[4745]: I1010 13:30:16.267235 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/128069b5-6f89-4da5-9e35-bfe14f1ebd50-bundle\") pod \"73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc\" (UID: \"128069b5-6f89-4da5-9e35-bfe14f1ebd50\") " pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" Oct 10 13:30:16 crc kubenswrapper[4745]: I1010 13:30:16.267267 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/128069b5-6f89-4da5-9e35-bfe14f1ebd50-util\") pod \"73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc\" (UID: \"128069b5-6f89-4da5-9e35-bfe14f1ebd50\") " pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" Oct 10 13:30:16 crc kubenswrapper[4745]: I1010 13:30:16.267666 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/128069b5-6f89-4da5-9e35-bfe14f1ebd50-util\") pod \"73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc\" (UID: \"128069b5-6f89-4da5-9e35-bfe14f1ebd50\") " pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" Oct 10 13:30:16 crc kubenswrapper[4745]: I1010 13:30:16.267823 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/128069b5-6f89-4da5-9e35-bfe14f1ebd50-bundle\") pod \"73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc\" (UID: \"128069b5-6f89-4da5-9e35-bfe14f1ebd50\") " pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" Oct 10 13:30:16 crc kubenswrapper[4745]: I1010 13:30:16.290828 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b797b\" (UniqueName: \"kubernetes.io/projected/128069b5-6f89-4da5-9e35-bfe14f1ebd50-kube-api-access-b797b\") pod \"73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc\" (UID: \"128069b5-6f89-4da5-9e35-bfe14f1ebd50\") " pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" Oct 10 13:30:16 crc kubenswrapper[4745]: I1010 13:30:16.424152 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" Oct 10 13:30:16 crc kubenswrapper[4745]: I1010 13:30:16.873882 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc"] Oct 10 13:30:16 crc kubenswrapper[4745]: W1010 13:30:16.881453 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod128069b5_6f89_4da5_9e35_bfe14f1ebd50.slice/crio-cbd498c5fd82c5ee53592f96ad55924574812f52e1dd19fa178e2d9b977a0523 WatchSource:0}: Error finding container cbd498c5fd82c5ee53592f96ad55924574812f52e1dd19fa178e2d9b977a0523: Status 404 returned error can't find the container with id cbd498c5fd82c5ee53592f96ad55924574812f52e1dd19fa178e2d9b977a0523 Oct 10 13:30:17 crc kubenswrapper[4745]: I1010 13:30:17.161153 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" event={"ID":"128069b5-6f89-4da5-9e35-bfe14f1ebd50","Type":"ContainerStarted","Data":"f67de0a6cb8432dd9ac8ace826a99274b6013fd0bc7f7412b903a877380215e2"} Oct 10 13:30:17 crc kubenswrapper[4745]: I1010 13:30:17.161199 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" event={"ID":"128069b5-6f89-4da5-9e35-bfe14f1ebd50","Type":"ContainerStarted","Data":"cbd498c5fd82c5ee53592f96ad55924574812f52e1dd19fa178e2d9b977a0523"} Oct 10 13:30:18 crc kubenswrapper[4745]: I1010 13:30:18.169674 4745 generic.go:334] "Generic (PLEG): container finished" podID="128069b5-6f89-4da5-9e35-bfe14f1ebd50" containerID="f67de0a6cb8432dd9ac8ace826a99274b6013fd0bc7f7412b903a877380215e2" exitCode=0 Oct 10 13:30:18 crc kubenswrapper[4745]: I1010 13:30:18.169761 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" event={"ID":"128069b5-6f89-4da5-9e35-bfe14f1ebd50","Type":"ContainerDied","Data":"f67de0a6cb8432dd9ac8ace826a99274b6013fd0bc7f7412b903a877380215e2"} Oct 10 13:30:19 crc kubenswrapper[4745]: I1010 13:30:19.182347 4745 generic.go:334] "Generic (PLEG): container finished" podID="128069b5-6f89-4da5-9e35-bfe14f1ebd50" containerID="e3b4e34f3991406eab58d850626e18b5fe31010331d0cd9ffbcf31c4c5570069" exitCode=0 Oct 10 13:30:19 crc kubenswrapper[4745]: I1010 13:30:19.182431 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" event={"ID":"128069b5-6f89-4da5-9e35-bfe14f1ebd50","Type":"ContainerDied","Data":"e3b4e34f3991406eab58d850626e18b5fe31010331d0cd9ffbcf31c4c5570069"} Oct 10 13:30:20 crc kubenswrapper[4745]: I1010 13:30:20.193118 4745 generic.go:334] "Generic (PLEG): container finished" podID="128069b5-6f89-4da5-9e35-bfe14f1ebd50" containerID="43dcf47186ab4d64f19fb7834c47e874ada94e5c3c221d2b56d5c6acdfbfd984" exitCode=0 Oct 10 13:30:20 crc kubenswrapper[4745]: I1010 13:30:20.193171 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" event={"ID":"128069b5-6f89-4da5-9e35-bfe14f1ebd50","Type":"ContainerDied","Data":"43dcf47186ab4d64f19fb7834c47e874ada94e5c3c221d2b56d5c6acdfbfd984"} Oct 10 13:30:21 crc kubenswrapper[4745]: I1010 13:30:21.613075 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" Oct 10 13:30:21 crc kubenswrapper[4745]: I1010 13:30:21.743969 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b797b\" (UniqueName: \"kubernetes.io/projected/128069b5-6f89-4da5-9e35-bfe14f1ebd50-kube-api-access-b797b\") pod \"128069b5-6f89-4da5-9e35-bfe14f1ebd50\" (UID: \"128069b5-6f89-4da5-9e35-bfe14f1ebd50\") " Oct 10 13:30:21 crc kubenswrapper[4745]: I1010 13:30:21.744205 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/128069b5-6f89-4da5-9e35-bfe14f1ebd50-util\") pod \"128069b5-6f89-4da5-9e35-bfe14f1ebd50\" (UID: \"128069b5-6f89-4da5-9e35-bfe14f1ebd50\") " Oct 10 13:30:21 crc kubenswrapper[4745]: I1010 13:30:21.744287 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/128069b5-6f89-4da5-9e35-bfe14f1ebd50-bundle\") pod \"128069b5-6f89-4da5-9e35-bfe14f1ebd50\" (UID: \"128069b5-6f89-4da5-9e35-bfe14f1ebd50\") " Oct 10 13:30:21 crc kubenswrapper[4745]: I1010 13:30:21.745316 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/128069b5-6f89-4da5-9e35-bfe14f1ebd50-bundle" (OuterVolumeSpecName: "bundle") pod "128069b5-6f89-4da5-9e35-bfe14f1ebd50" (UID: "128069b5-6f89-4da5-9e35-bfe14f1ebd50"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:30:21 crc kubenswrapper[4745]: I1010 13:30:21.752319 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/128069b5-6f89-4da5-9e35-bfe14f1ebd50-kube-api-access-b797b" (OuterVolumeSpecName: "kube-api-access-b797b") pod "128069b5-6f89-4da5-9e35-bfe14f1ebd50" (UID: "128069b5-6f89-4da5-9e35-bfe14f1ebd50"). InnerVolumeSpecName "kube-api-access-b797b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:30:21 crc kubenswrapper[4745]: I1010 13:30:21.771383 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/128069b5-6f89-4da5-9e35-bfe14f1ebd50-util" (OuterVolumeSpecName: "util") pod "128069b5-6f89-4da5-9e35-bfe14f1ebd50" (UID: "128069b5-6f89-4da5-9e35-bfe14f1ebd50"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:30:21 crc kubenswrapper[4745]: I1010 13:30:21.845690 4745 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/128069b5-6f89-4da5-9e35-bfe14f1ebd50-util\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:21 crc kubenswrapper[4745]: I1010 13:30:21.845756 4745 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/128069b5-6f89-4da5-9e35-bfe14f1ebd50-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:21 crc kubenswrapper[4745]: I1010 13:30:21.845774 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b797b\" (UniqueName: \"kubernetes.io/projected/128069b5-6f89-4da5-9e35-bfe14f1ebd50-kube-api-access-b797b\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:22 crc kubenswrapper[4745]: I1010 13:30:22.214812 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" event={"ID":"128069b5-6f89-4da5-9e35-bfe14f1ebd50","Type":"ContainerDied","Data":"cbd498c5fd82c5ee53592f96ad55924574812f52e1dd19fa178e2d9b977a0523"} Oct 10 13:30:22 crc kubenswrapper[4745]: I1010 13:30:22.214875 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbd498c5fd82c5ee53592f96ad55924574812f52e1dd19fa178e2d9b977a0523" Oct 10 13:30:22 crc kubenswrapper[4745]: I1010 13:30:22.214923 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc" Oct 10 13:30:28 crc kubenswrapper[4745]: I1010 13:30:28.895295 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7f65679b66-d82sn"] Oct 10 13:30:28 crc kubenswrapper[4745]: E1010 13:30:28.895947 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="128069b5-6f89-4da5-9e35-bfe14f1ebd50" containerName="util" Oct 10 13:30:28 crc kubenswrapper[4745]: I1010 13:30:28.895962 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="128069b5-6f89-4da5-9e35-bfe14f1ebd50" containerName="util" Oct 10 13:30:28 crc kubenswrapper[4745]: E1010 13:30:28.895974 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="128069b5-6f89-4da5-9e35-bfe14f1ebd50" containerName="extract" Oct 10 13:30:28 crc kubenswrapper[4745]: I1010 13:30:28.895981 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="128069b5-6f89-4da5-9e35-bfe14f1ebd50" containerName="extract" Oct 10 13:30:28 crc kubenswrapper[4745]: E1010 13:30:28.895998 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="128069b5-6f89-4da5-9e35-bfe14f1ebd50" containerName="pull" Oct 10 13:30:28 crc kubenswrapper[4745]: I1010 13:30:28.896006 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="128069b5-6f89-4da5-9e35-bfe14f1ebd50" containerName="pull" Oct 10 13:30:28 crc kubenswrapper[4745]: I1010 13:30:28.896149 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="128069b5-6f89-4da5-9e35-bfe14f1ebd50" containerName="extract" Oct 10 13:30:28 crc kubenswrapper[4745]: I1010 13:30:28.896818 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7f65679b66-d82sn" Oct 10 13:30:28 crc kubenswrapper[4745]: I1010 13:30:28.901246 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-h8tfl" Oct 10 13:30:28 crc kubenswrapper[4745]: I1010 13:30:28.920619 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7f65679b66-d82sn"] Oct 10 13:30:28 crc kubenswrapper[4745]: I1010 13:30:28.963397 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4q7j\" (UniqueName: \"kubernetes.io/projected/b0e08ed6-f210-41b7-907e-d6722025f43e-kube-api-access-k4q7j\") pod \"openstack-operator-controller-operator-7f65679b66-d82sn\" (UID: \"b0e08ed6-f210-41b7-907e-d6722025f43e\") " pod="openstack-operators/openstack-operator-controller-operator-7f65679b66-d82sn" Oct 10 13:30:29 crc kubenswrapper[4745]: I1010 13:30:29.064422 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4q7j\" (UniqueName: \"kubernetes.io/projected/b0e08ed6-f210-41b7-907e-d6722025f43e-kube-api-access-k4q7j\") pod \"openstack-operator-controller-operator-7f65679b66-d82sn\" (UID: \"b0e08ed6-f210-41b7-907e-d6722025f43e\") " pod="openstack-operators/openstack-operator-controller-operator-7f65679b66-d82sn" Oct 10 13:30:29 crc kubenswrapper[4745]: I1010 13:30:29.084549 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4q7j\" (UniqueName: \"kubernetes.io/projected/b0e08ed6-f210-41b7-907e-d6722025f43e-kube-api-access-k4q7j\") pod \"openstack-operator-controller-operator-7f65679b66-d82sn\" (UID: \"b0e08ed6-f210-41b7-907e-d6722025f43e\") " pod="openstack-operators/openstack-operator-controller-operator-7f65679b66-d82sn" Oct 10 13:30:29 crc kubenswrapper[4745]: I1010 13:30:29.230913 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7f65679b66-d82sn" Oct 10 13:30:29 crc kubenswrapper[4745]: I1010 13:30:29.722091 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7f65679b66-d82sn"] Oct 10 13:30:29 crc kubenswrapper[4745]: W1010 13:30:29.737180 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0e08ed6_f210_41b7_907e_d6722025f43e.slice/crio-487693fb6d96113c808f58aa882785ae91395b52ca1110c6a5c5e2c6fbc32490 WatchSource:0}: Error finding container 487693fb6d96113c808f58aa882785ae91395b52ca1110c6a5c5e2c6fbc32490: Status 404 returned error can't find the container with id 487693fb6d96113c808f58aa882785ae91395b52ca1110c6a5c5e2c6fbc32490 Oct 10 13:30:30 crc kubenswrapper[4745]: I1010 13:30:30.317157 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7f65679b66-d82sn" event={"ID":"b0e08ed6-f210-41b7-907e-d6722025f43e","Type":"ContainerStarted","Data":"487693fb6d96113c808f58aa882785ae91395b52ca1110c6a5c5e2c6fbc32490"} Oct 10 13:30:33 crc kubenswrapper[4745]: I1010 13:30:33.338309 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7f65679b66-d82sn" event={"ID":"b0e08ed6-f210-41b7-907e-d6722025f43e","Type":"ContainerStarted","Data":"c0e98650ef7d9cfd000f480ac7438e4c0edab20811fc3721bbbc85dc6bca53f6"} Oct 10 13:30:36 crc kubenswrapper[4745]: I1010 13:30:36.355204 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7f65679b66-d82sn" event={"ID":"b0e08ed6-f210-41b7-907e-d6722025f43e","Type":"ContainerStarted","Data":"f88206b8e472a896a934d69ac0e1e7ff970b3d2a01f1fef7b0e01fda8979f8b0"} Oct 10 13:30:36 crc kubenswrapper[4745]: I1010 13:30:36.355649 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7f65679b66-d82sn" Oct 10 13:30:36 crc kubenswrapper[4745]: I1010 13:30:36.385484 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7f65679b66-d82sn" podStartSLOduration=2.889209083 podStartE2EDuration="8.38547098s" podCreationTimestamp="2025-10-10 13:30:28 +0000 UTC" firstStartedPulling="2025-10-10 13:30:29.74078725 +0000 UTC m=+743.638444023" lastFinishedPulling="2025-10-10 13:30:35.237049157 +0000 UTC m=+749.134705920" observedRunningTime="2025-10-10 13:30:36.381857177 +0000 UTC m=+750.279513940" watchObservedRunningTime="2025-10-10 13:30:36.38547098 +0000 UTC m=+750.283127743" Oct 10 13:30:39 crc kubenswrapper[4745]: I1010 13:30:39.234880 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7f65679b66-d82sn" Oct 10 13:30:40 crc kubenswrapper[4745]: I1010 13:30:40.422792 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7g9x6"] Oct 10 13:30:40 crc kubenswrapper[4745]: I1010 13:30:40.424193 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7g9x6" Oct 10 13:30:40 crc kubenswrapper[4745]: I1010 13:30:40.444474 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7g9x6"] Oct 10 13:30:40 crc kubenswrapper[4745]: I1010 13:30:40.542443 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd7d5df-eee1-4f07-8dd8-24873ec41827-catalog-content\") pod \"redhat-operators-7g9x6\" (UID: \"2dd7d5df-eee1-4f07-8dd8-24873ec41827\") " pod="openshift-marketplace/redhat-operators-7g9x6" Oct 10 13:30:40 crc kubenswrapper[4745]: I1010 13:30:40.542519 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gnvh\" (UniqueName: \"kubernetes.io/projected/2dd7d5df-eee1-4f07-8dd8-24873ec41827-kube-api-access-6gnvh\") pod \"redhat-operators-7g9x6\" (UID: \"2dd7d5df-eee1-4f07-8dd8-24873ec41827\") " pod="openshift-marketplace/redhat-operators-7g9x6" Oct 10 13:30:40 crc kubenswrapper[4745]: I1010 13:30:40.542549 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd7d5df-eee1-4f07-8dd8-24873ec41827-utilities\") pod \"redhat-operators-7g9x6\" (UID: \"2dd7d5df-eee1-4f07-8dd8-24873ec41827\") " pod="openshift-marketplace/redhat-operators-7g9x6" Oct 10 13:30:40 crc kubenswrapper[4745]: I1010 13:30:40.643647 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd7d5df-eee1-4f07-8dd8-24873ec41827-catalog-content\") pod \"redhat-operators-7g9x6\" (UID: \"2dd7d5df-eee1-4f07-8dd8-24873ec41827\") " pod="openshift-marketplace/redhat-operators-7g9x6" Oct 10 13:30:40 crc kubenswrapper[4745]: I1010 13:30:40.643722 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gnvh\" (UniqueName: \"kubernetes.io/projected/2dd7d5df-eee1-4f07-8dd8-24873ec41827-kube-api-access-6gnvh\") pod \"redhat-operators-7g9x6\" (UID: \"2dd7d5df-eee1-4f07-8dd8-24873ec41827\") " pod="openshift-marketplace/redhat-operators-7g9x6" Oct 10 13:30:40 crc kubenswrapper[4745]: I1010 13:30:40.643788 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd7d5df-eee1-4f07-8dd8-24873ec41827-utilities\") pod \"redhat-operators-7g9x6\" (UID: \"2dd7d5df-eee1-4f07-8dd8-24873ec41827\") " pod="openshift-marketplace/redhat-operators-7g9x6" Oct 10 13:30:40 crc kubenswrapper[4745]: I1010 13:30:40.644454 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd7d5df-eee1-4f07-8dd8-24873ec41827-utilities\") pod \"redhat-operators-7g9x6\" (UID: \"2dd7d5df-eee1-4f07-8dd8-24873ec41827\") " pod="openshift-marketplace/redhat-operators-7g9x6" Oct 10 13:30:40 crc kubenswrapper[4745]: I1010 13:30:40.644455 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd7d5df-eee1-4f07-8dd8-24873ec41827-catalog-content\") pod \"redhat-operators-7g9x6\" (UID: \"2dd7d5df-eee1-4f07-8dd8-24873ec41827\") " pod="openshift-marketplace/redhat-operators-7g9x6" Oct 10 13:30:40 crc kubenswrapper[4745]: I1010 13:30:40.670074 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gnvh\" (UniqueName: \"kubernetes.io/projected/2dd7d5df-eee1-4f07-8dd8-24873ec41827-kube-api-access-6gnvh\") pod \"redhat-operators-7g9x6\" (UID: \"2dd7d5df-eee1-4f07-8dd8-24873ec41827\") " pod="openshift-marketplace/redhat-operators-7g9x6" Oct 10 13:30:40 crc kubenswrapper[4745]: I1010 13:30:40.789237 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7g9x6" Oct 10 13:30:41 crc kubenswrapper[4745]: I1010 13:30:41.207267 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7g9x6"] Oct 10 13:30:41 crc kubenswrapper[4745]: I1010 13:30:41.394347 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7g9x6" event={"ID":"2dd7d5df-eee1-4f07-8dd8-24873ec41827","Type":"ContainerStarted","Data":"34df699941d4a668bc2b247cec10405629512ec51f65fdf8573b5513769d99cd"} Oct 10 13:30:41 crc kubenswrapper[4745]: I1010 13:30:41.394387 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7g9x6" event={"ID":"2dd7d5df-eee1-4f07-8dd8-24873ec41827","Type":"ContainerStarted","Data":"bbeefee00cb77bdf6f9cf1a8cdf816c680a63f5802ccb0d12067e6cb628ae1ff"} Oct 10 13:30:42 crc kubenswrapper[4745]: I1010 13:30:42.400835 4745 generic.go:334] "Generic (PLEG): container finished" podID="2dd7d5df-eee1-4f07-8dd8-24873ec41827" containerID="34df699941d4a668bc2b247cec10405629512ec51f65fdf8573b5513769d99cd" exitCode=0 Oct 10 13:30:42 crc kubenswrapper[4745]: I1010 13:30:42.400899 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7g9x6" event={"ID":"2dd7d5df-eee1-4f07-8dd8-24873ec41827","Type":"ContainerDied","Data":"34df699941d4a668bc2b247cec10405629512ec51f65fdf8573b5513769d99cd"} Oct 10 13:30:43 crc kubenswrapper[4745]: I1010 13:30:43.406970 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7g9x6" event={"ID":"2dd7d5df-eee1-4f07-8dd8-24873ec41827","Type":"ContainerStarted","Data":"52d5539698174948775e905fdf84bbc76aff1b853796e85220df5bb3264ee9ba"} Oct 10 13:30:44 crc kubenswrapper[4745]: I1010 13:30:44.414271 4745 generic.go:334] "Generic (PLEG): container finished" podID="2dd7d5df-eee1-4f07-8dd8-24873ec41827" containerID="52d5539698174948775e905fdf84bbc76aff1b853796e85220df5bb3264ee9ba" exitCode=0 Oct 10 13:30:44 crc kubenswrapper[4745]: I1010 13:30:44.414329 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7g9x6" event={"ID":"2dd7d5df-eee1-4f07-8dd8-24873ec41827","Type":"ContainerDied","Data":"52d5539698174948775e905fdf84bbc76aff1b853796e85220df5bb3264ee9ba"} Oct 10 13:30:45 crc kubenswrapper[4745]: I1010 13:30:45.425756 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7g9x6" event={"ID":"2dd7d5df-eee1-4f07-8dd8-24873ec41827","Type":"ContainerStarted","Data":"4c35453f0d240185f0b98fafd3e9fde2ea6ed29883dfa2109c3f746a1964afda"} Oct 10 13:30:45 crc kubenswrapper[4745]: I1010 13:30:45.454960 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7g9x6" podStartSLOduration=2.9851882400000003 podStartE2EDuration="5.454944943s" podCreationTimestamp="2025-10-10 13:30:40 +0000 UTC" firstStartedPulling="2025-10-10 13:30:42.403602119 +0000 UTC m=+756.301258882" lastFinishedPulling="2025-10-10 13:30:44.873358822 +0000 UTC m=+758.771015585" observedRunningTime="2025-10-10 13:30:45.45438799 +0000 UTC m=+759.352044793" watchObservedRunningTime="2025-10-10 13:30:45.454944943 +0000 UTC m=+759.352601706" Oct 10 13:30:46 crc kubenswrapper[4745]: I1010 13:30:46.186677 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:30:46 crc kubenswrapper[4745]: I1010 13:30:46.187045 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:30:46 crc kubenswrapper[4745]: I1010 13:30:46.187107 4745 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:30:46 crc kubenswrapper[4745]: I1010 13:30:46.187917 4745 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2ba538b1feacb1e3fe2ff42d8dddb15b3e992a880af6f43548cdf890834090dc"} pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 13:30:46 crc kubenswrapper[4745]: I1010 13:30:46.188022 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" containerID="cri-o://2ba538b1feacb1e3fe2ff42d8dddb15b3e992a880af6f43548cdf890834090dc" gracePeriod=600 Oct 10 13:30:46 crc kubenswrapper[4745]: I1010 13:30:46.441004 4745 generic.go:334] "Generic (PLEG): container finished" podID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerID="2ba538b1feacb1e3fe2ff42d8dddb15b3e992a880af6f43548cdf890834090dc" exitCode=0 Oct 10 13:30:46 crc kubenswrapper[4745]: I1010 13:30:46.441088 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerDied","Data":"2ba538b1feacb1e3fe2ff42d8dddb15b3e992a880af6f43548cdf890834090dc"} Oct 10 13:30:46 crc kubenswrapper[4745]: I1010 13:30:46.441145 4745 scope.go:117] "RemoveContainer" containerID="c5437d8915ec61a5a5bd016aa1daf75ecd5d17ec75528801c9e54e26c6856701" Oct 10 13:30:47 crc kubenswrapper[4745]: I1010 13:30:47.448844 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"8a31c6076bf04e3e0b289d6ba4cb6b6d1eeb3e2ba416958cc06b859417161f1e"} Oct 10 13:30:47 crc kubenswrapper[4745]: I1010 13:30:47.608149 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9jkxn"] Oct 10 13:30:47 crc kubenswrapper[4745]: I1010 13:30:47.609871 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9jkxn" Oct 10 13:30:47 crc kubenswrapper[4745]: I1010 13:30:47.629014 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9jkxn"] Oct 10 13:30:47 crc kubenswrapper[4745]: I1010 13:30:47.643189 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e605d21-bc5f-44ee-a3f7-7558d307162b-utilities\") pod \"redhat-marketplace-9jkxn\" (UID: \"8e605d21-bc5f-44ee-a3f7-7558d307162b\") " pod="openshift-marketplace/redhat-marketplace-9jkxn" Oct 10 13:30:47 crc kubenswrapper[4745]: I1010 13:30:47.643284 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e605d21-bc5f-44ee-a3f7-7558d307162b-catalog-content\") pod \"redhat-marketplace-9jkxn\" (UID: \"8e605d21-bc5f-44ee-a3f7-7558d307162b\") " pod="openshift-marketplace/redhat-marketplace-9jkxn" Oct 10 13:30:47 crc kubenswrapper[4745]: I1010 13:30:47.643424 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxn5z\" (UniqueName: \"kubernetes.io/projected/8e605d21-bc5f-44ee-a3f7-7558d307162b-kube-api-access-rxn5z\") pod \"redhat-marketplace-9jkxn\" (UID: \"8e605d21-bc5f-44ee-a3f7-7558d307162b\") " pod="openshift-marketplace/redhat-marketplace-9jkxn" Oct 10 13:30:47 crc kubenswrapper[4745]: I1010 13:30:47.744875 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxn5z\" (UniqueName: \"kubernetes.io/projected/8e605d21-bc5f-44ee-a3f7-7558d307162b-kube-api-access-rxn5z\") pod \"redhat-marketplace-9jkxn\" (UID: \"8e605d21-bc5f-44ee-a3f7-7558d307162b\") " pod="openshift-marketplace/redhat-marketplace-9jkxn" Oct 10 13:30:47 crc kubenswrapper[4745]: I1010 13:30:47.744946 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e605d21-bc5f-44ee-a3f7-7558d307162b-utilities\") pod \"redhat-marketplace-9jkxn\" (UID: \"8e605d21-bc5f-44ee-a3f7-7558d307162b\") " pod="openshift-marketplace/redhat-marketplace-9jkxn" Oct 10 13:30:47 crc kubenswrapper[4745]: I1010 13:30:47.744981 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e605d21-bc5f-44ee-a3f7-7558d307162b-catalog-content\") pod \"redhat-marketplace-9jkxn\" (UID: \"8e605d21-bc5f-44ee-a3f7-7558d307162b\") " pod="openshift-marketplace/redhat-marketplace-9jkxn" Oct 10 13:30:47 crc kubenswrapper[4745]: I1010 13:30:47.745451 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e605d21-bc5f-44ee-a3f7-7558d307162b-catalog-content\") pod \"redhat-marketplace-9jkxn\" (UID: \"8e605d21-bc5f-44ee-a3f7-7558d307162b\") " pod="openshift-marketplace/redhat-marketplace-9jkxn" Oct 10 13:30:47 crc kubenswrapper[4745]: I1010 13:30:47.745758 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e605d21-bc5f-44ee-a3f7-7558d307162b-utilities\") pod \"redhat-marketplace-9jkxn\" (UID: \"8e605d21-bc5f-44ee-a3f7-7558d307162b\") " pod="openshift-marketplace/redhat-marketplace-9jkxn" Oct 10 13:30:47 crc kubenswrapper[4745]: I1010 13:30:47.772289 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxn5z\" (UniqueName: \"kubernetes.io/projected/8e605d21-bc5f-44ee-a3f7-7558d307162b-kube-api-access-rxn5z\") pod \"redhat-marketplace-9jkxn\" (UID: \"8e605d21-bc5f-44ee-a3f7-7558d307162b\") " pod="openshift-marketplace/redhat-marketplace-9jkxn" Oct 10 13:30:47 crc kubenswrapper[4745]: I1010 13:30:47.940933 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9jkxn" Oct 10 13:30:48 crc kubenswrapper[4745]: I1010 13:30:48.383013 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9jkxn"] Oct 10 13:30:48 crc kubenswrapper[4745]: W1010 13:30:48.389918 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e605d21_bc5f_44ee_a3f7_7558d307162b.slice/crio-0f659751771f20c1317b8363985766299f9773610eac060c31094c9f8aa40ca7 WatchSource:0}: Error finding container 0f659751771f20c1317b8363985766299f9773610eac060c31094c9f8aa40ca7: Status 404 returned error can't find the container with id 0f659751771f20c1317b8363985766299f9773610eac060c31094c9f8aa40ca7 Oct 10 13:30:48 crc kubenswrapper[4745]: I1010 13:30:48.454428 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jkxn" event={"ID":"8e605d21-bc5f-44ee-a3f7-7558d307162b","Type":"ContainerStarted","Data":"0f659751771f20c1317b8363985766299f9773610eac060c31094c9f8aa40ca7"} Oct 10 13:30:49 crc kubenswrapper[4745]: I1010 13:30:49.461628 4745 generic.go:334] "Generic (PLEG): container finished" podID="8e605d21-bc5f-44ee-a3f7-7558d307162b" containerID="856d6a9fa1720c9f341eb44b43ad53ec1793fbbc7d6e925b18a0fd00788c7aa8" exitCode=0 Oct 10 13:30:49 crc kubenswrapper[4745]: I1010 13:30:49.461676 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jkxn" event={"ID":"8e605d21-bc5f-44ee-a3f7-7558d307162b","Type":"ContainerDied","Data":"856d6a9fa1720c9f341eb44b43ad53ec1793fbbc7d6e925b18a0fd00788c7aa8"} Oct 10 13:30:50 crc kubenswrapper[4745]: I1010 13:30:50.789668 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7g9x6" Oct 10 13:30:50 crc kubenswrapper[4745]: I1010 13:30:50.791871 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7g9x6" Oct 10 13:30:50 crc kubenswrapper[4745]: I1010 13:30:50.852079 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7g9x6" Oct 10 13:30:51 crc kubenswrapper[4745]: I1010 13:30:51.476565 4745 generic.go:334] "Generic (PLEG): container finished" podID="8e605d21-bc5f-44ee-a3f7-7558d307162b" containerID="de2ca6f748b81cd399df67160f038b7a52277e5e3b0efb0b149cd3ee230681bc" exitCode=0 Oct 10 13:30:51 crc kubenswrapper[4745]: I1010 13:30:51.476685 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jkxn" event={"ID":"8e605d21-bc5f-44ee-a3f7-7558d307162b","Type":"ContainerDied","Data":"de2ca6f748b81cd399df67160f038b7a52277e5e3b0efb0b149cd3ee230681bc"} Oct 10 13:30:51 crc kubenswrapper[4745]: I1010 13:30:51.530066 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7g9x6" Oct 10 13:30:52 crc kubenswrapper[4745]: I1010 13:30:52.401367 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7g9x6"] Oct 10 13:30:52 crc kubenswrapper[4745]: I1010 13:30:52.487222 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jkxn" event={"ID":"8e605d21-bc5f-44ee-a3f7-7558d307162b","Type":"ContainerStarted","Data":"136dcf058e7d5d64edee0c76f38356fb2bbf311dcc3d135959ac715111cfa951"} Oct 10 13:30:52 crc kubenswrapper[4745]: I1010 13:30:52.510911 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9jkxn" podStartSLOduration=3.013316806 podStartE2EDuration="5.510896357s" podCreationTimestamp="2025-10-10 13:30:47 +0000 UTC" firstStartedPulling="2025-10-10 13:30:49.463518484 +0000 UTC m=+763.361175247" lastFinishedPulling="2025-10-10 13:30:51.961098035 +0000 UTC m=+765.858754798" observedRunningTime="2025-10-10 13:30:52.508688116 +0000 UTC m=+766.406344889" watchObservedRunningTime="2025-10-10 13:30:52.510896357 +0000 UTC m=+766.408553120" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.212712 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-htmcv"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.214178 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-htmcv" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.221925 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-mrnwj" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.235059 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-htmcv"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.247427 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-c7986"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.248633 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-c7986" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.252354 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-kxmcq" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.260334 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-c7986"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.263845 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-fpcxk"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.264990 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-fpcxk" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.272188 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-fpcxk"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.286088 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-nd8l6" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.288193 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-tz665"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.289203 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tz665" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.292207 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-vfdxt" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.295760 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-2w6nw"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.298514 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-2w6nw" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.300887 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-x5gjl" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.318796 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-tz665"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.335451 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-2w6nw"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.337268 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs54k\" (UniqueName: \"kubernetes.io/projected/730b5555-5f6d-4e16-91e0-faacad56d777-kube-api-access-hs54k\") pod \"barbican-operator-controller-manager-64f84fcdbb-htmcv\" (UID: \"730b5555-5f6d-4e16-91e0-faacad56d777\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-htmcv" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.337385 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcxtk\" (UniqueName: \"kubernetes.io/projected/446e0b63-81e5-47fb-b110-cfa5393e5311-kube-api-access-mcxtk\") pod \"glance-operator-controller-manager-7bb46cd7d-tz665\" (UID: \"446e0b63-81e5-47fb-b110-cfa5393e5311\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tz665" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.337457 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc4rk\" (UniqueName: \"kubernetes.io/projected/5ca519af-249e-48f0-8669-77bd548be8f3-kube-api-access-qc4rk\") pod \"cinder-operator-controller-manager-59cdc64769-c7986\" (UID: \"5ca519af-249e-48f0-8669-77bd548be8f3\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-c7986" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.337609 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dg7g\" (UniqueName: \"kubernetes.io/projected/dff847b8-915b-4b64-85c3-d7c20d1282a0-kube-api-access-4dg7g\") pod \"designate-operator-controller-manager-687df44cdb-fpcxk\" (UID: \"dff847b8-915b-4b64-85c3-d7c20d1282a0\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-fpcxk" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.374841 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-qmbvh"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.376232 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-qmbvh" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.378208 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-5z88w" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.394623 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-qmbvh"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.409960 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.410985 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.412983 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.413195 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-wk8kn" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.430339 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-d5qnn"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.431852 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-d5qnn" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.438956 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-c72nj" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.439820 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w97xd\" (UniqueName: \"kubernetes.io/projected/7f07678c-0cc2-47a5-b244-72ae3aa55baa-kube-api-access-w97xd\") pod \"heat-operator-controller-manager-6d9967f8dd-2w6nw\" (UID: \"7f07678c-0cc2-47a5-b244-72ae3aa55baa\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-2w6nw" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.439874 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc4rk\" (UniqueName: \"kubernetes.io/projected/5ca519af-249e-48f0-8669-77bd548be8f3-kube-api-access-qc4rk\") pod \"cinder-operator-controller-manager-59cdc64769-c7986\" (UID: \"5ca519af-249e-48f0-8669-77bd548be8f3\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-c7986" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.439901 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dg7g\" (UniqueName: \"kubernetes.io/projected/dff847b8-915b-4b64-85c3-d7c20d1282a0-kube-api-access-4dg7g\") pod \"designate-operator-controller-manager-687df44cdb-fpcxk\" (UID: \"dff847b8-915b-4b64-85c3-d7c20d1282a0\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-fpcxk" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.439927 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs54k\" (UniqueName: \"kubernetes.io/projected/730b5555-5f6d-4e16-91e0-faacad56d777-kube-api-access-hs54k\") pod \"barbican-operator-controller-manager-64f84fcdbb-htmcv\" (UID: \"730b5555-5f6d-4e16-91e0-faacad56d777\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-htmcv" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.439982 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dhks\" (UniqueName: \"kubernetes.io/projected/1d2b124e-5495-42bd-b675-4eed70ccc446-kube-api-access-7dhks\") pod \"horizon-operator-controller-manager-6d74794d9b-qmbvh\" (UID: \"1d2b124e-5495-42bd-b675-4eed70ccc446\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-qmbvh" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.440039 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcxtk\" (UniqueName: \"kubernetes.io/projected/446e0b63-81e5-47fb-b110-cfa5393e5311-kube-api-access-mcxtk\") pod \"glance-operator-controller-manager-7bb46cd7d-tz665\" (UID: \"446e0b63-81e5-47fb-b110-cfa5393e5311\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tz665" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.444639 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.465017 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc4rk\" (UniqueName: \"kubernetes.io/projected/5ca519af-249e-48f0-8669-77bd548be8f3-kube-api-access-qc4rk\") pod \"cinder-operator-controller-manager-59cdc64769-c7986\" (UID: \"5ca519af-249e-48f0-8669-77bd548be8f3\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-c7986" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.465715 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs54k\" (UniqueName: \"kubernetes.io/projected/730b5555-5f6d-4e16-91e0-faacad56d777-kube-api-access-hs54k\") pod \"barbican-operator-controller-manager-64f84fcdbb-htmcv\" (UID: \"730b5555-5f6d-4e16-91e0-faacad56d777\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-htmcv" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.468524 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcxtk\" (UniqueName: \"kubernetes.io/projected/446e0b63-81e5-47fb-b110-cfa5393e5311-kube-api-access-mcxtk\") pod \"glance-operator-controller-manager-7bb46cd7d-tz665\" (UID: \"446e0b63-81e5-47fb-b110-cfa5393e5311\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tz665" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.470144 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-h2bt5"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.471703 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-h2bt5" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.477629 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-d5qnn"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.478480 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dg7g\" (UniqueName: \"kubernetes.io/projected/dff847b8-915b-4b64-85c3-d7c20d1282a0-kube-api-access-4dg7g\") pod \"designate-operator-controller-manager-687df44cdb-fpcxk\" (UID: \"dff847b8-915b-4b64-85c3-d7c20d1282a0\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-fpcxk" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.483105 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-r56d9" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.491640 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-t9wqw"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.492597 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-t9wqw" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.501323 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-n8f8g" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.509184 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7g9x6" podUID="2dd7d5df-eee1-4f07-8dd8-24873ec41827" containerName="registry-server" containerID="cri-o://4c35453f0d240185f0b98fafd3e9fde2ea6ed29883dfa2109c3f746a1964afda" gracePeriod=2 Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.518942 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-vc9rk"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.520064 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-vc9rk" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.522397 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-h75gt" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.529571 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-htmcv" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.536671 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-h2bt5"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.542681 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/443416f7-e4ce-41f2-b546-9344f0cb72a7-cert\") pod \"infra-operator-controller-manager-585fc5b659-dt9kx\" (UID: \"443416f7-e4ce-41f2-b546-9344f0cb72a7\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.542724 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dhks\" (UniqueName: \"kubernetes.io/projected/1d2b124e-5495-42bd-b675-4eed70ccc446-kube-api-access-7dhks\") pod \"horizon-operator-controller-manager-6d74794d9b-qmbvh\" (UID: \"1d2b124e-5495-42bd-b675-4eed70ccc446\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-qmbvh" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.542783 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjbw9\" (UniqueName: \"kubernetes.io/projected/443416f7-e4ce-41f2-b546-9344f0cb72a7-kube-api-access-kjbw9\") pod \"infra-operator-controller-manager-585fc5b659-dt9kx\" (UID: \"443416f7-e4ce-41f2-b546-9344f0cb72a7\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.542831 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2cx6\" (UniqueName: \"kubernetes.io/projected/fc401fc1-e03b-4b32-8782-15e9c2a70e8a-kube-api-access-v2cx6\") pod \"manila-operator-controller-manager-59578bc799-t9wqw\" (UID: \"fc401fc1-e03b-4b32-8782-15e9c2a70e8a\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-t9wqw" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.542865 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w97xd\" (UniqueName: \"kubernetes.io/projected/7f07678c-0cc2-47a5-b244-72ae3aa55baa-kube-api-access-w97xd\") pod \"heat-operator-controller-manager-6d9967f8dd-2w6nw\" (UID: \"7f07678c-0cc2-47a5-b244-72ae3aa55baa\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-2w6nw" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.542898 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h256t\" (UniqueName: \"kubernetes.io/projected/9fdf7ed4-1e18-43c0-b305-2b9b226f4634-kube-api-access-h256t\") pod \"ironic-operator-controller-manager-74cb5cbc49-d5qnn\" (UID: \"9fdf7ed4-1e18-43c0-b305-2b9b226f4634\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-d5qnn" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.542916 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6ndx\" (UniqueName: \"kubernetes.io/projected/10978ce4-0034-4ad4-9a47-2f1931c09655-kube-api-access-r6ndx\") pod \"keystone-operator-controller-manager-ddb98f99b-h2bt5\" (UID: \"10978ce4-0034-4ad4-9a47-2f1931c09655\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-h2bt5" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.549272 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-t9wqw"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.565548 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-c7986" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.567306 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-vc9rk"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.569652 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w97xd\" (UniqueName: \"kubernetes.io/projected/7f07678c-0cc2-47a5-b244-72ae3aa55baa-kube-api-access-w97xd\") pod \"heat-operator-controller-manager-6d9967f8dd-2w6nw\" (UID: \"7f07678c-0cc2-47a5-b244-72ae3aa55baa\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-2w6nw" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.569695 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-2dmpc"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.570585 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2dmpc" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.576811 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-ggw9m" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.577283 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dhks\" (UniqueName: \"kubernetes.io/projected/1d2b124e-5495-42bd-b675-4eed70ccc446-kube-api-access-7dhks\") pod \"horizon-operator-controller-manager-6d74794d9b-qmbvh\" (UID: \"1d2b124e-5495-42bd-b675-4eed70ccc446\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-qmbvh" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.588829 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-fpcxk" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.602860 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-89lzv"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.603897 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-89lzv" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.607247 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-fdthd" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.614620 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-2dmpc"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.617535 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tz665" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.629252 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-89lzv"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.643440 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-2w6nw" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.644198 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjbw9\" (UniqueName: \"kubernetes.io/projected/443416f7-e4ce-41f2-b546-9344f0cb72a7-kube-api-access-kjbw9\") pod \"infra-operator-controller-manager-585fc5b659-dt9kx\" (UID: \"443416f7-e4ce-41f2-b546-9344f0cb72a7\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.644297 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2cx6\" (UniqueName: \"kubernetes.io/projected/fc401fc1-e03b-4b32-8782-15e9c2a70e8a-kube-api-access-v2cx6\") pod \"manila-operator-controller-manager-59578bc799-t9wqw\" (UID: \"fc401fc1-e03b-4b32-8782-15e9c2a70e8a\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-t9wqw" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.644350 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b628\" (UniqueName: \"kubernetes.io/projected/83b0fe55-090f-448e-95c2-1d3cbbc1b54c-kube-api-access-9b628\") pod \"neutron-operator-controller-manager-797d478b46-2dmpc\" (UID: \"83b0fe55-090f-448e-95c2-1d3cbbc1b54c\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2dmpc" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.644990 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhngf\" (UniqueName: \"kubernetes.io/projected/420fdd11-0131-479d-9f7a-47f091cacd83-kube-api-access-nhngf\") pod \"mariadb-operator-controller-manager-5777b4f897-vc9rk\" (UID: \"420fdd11-0131-479d-9f7a-47f091cacd83\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-vc9rk" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.645058 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h256t\" (UniqueName: \"kubernetes.io/projected/9fdf7ed4-1e18-43c0-b305-2b9b226f4634-kube-api-access-h256t\") pod \"ironic-operator-controller-manager-74cb5cbc49-d5qnn\" (UID: \"9fdf7ed4-1e18-43c0-b305-2b9b226f4634\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-d5qnn" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.645086 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6ndx\" (UniqueName: \"kubernetes.io/projected/10978ce4-0034-4ad4-9a47-2f1931c09655-kube-api-access-r6ndx\") pod \"keystone-operator-controller-manager-ddb98f99b-h2bt5\" (UID: \"10978ce4-0034-4ad4-9a47-2f1931c09655\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-h2bt5" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.645167 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/443416f7-e4ce-41f2-b546-9344f0cb72a7-cert\") pod \"infra-operator-controller-manager-585fc5b659-dt9kx\" (UID: \"443416f7-e4ce-41f2-b546-9344f0cb72a7\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx" Oct 10 13:30:53 crc kubenswrapper[4745]: E1010 13:30:53.645856 4745 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 10 13:30:53 crc kubenswrapper[4745]: E1010 13:30:53.648008 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/443416f7-e4ce-41f2-b546-9344f0cb72a7-cert podName:443416f7-e4ce-41f2-b546-9344f0cb72a7 nodeName:}" failed. No retries permitted until 2025-10-10 13:30:54.147983211 +0000 UTC m=+768.045639974 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/443416f7-e4ce-41f2-b546-9344f0cb72a7-cert") pod "infra-operator-controller-manager-585fc5b659-dt9kx" (UID: "443416f7-e4ce-41f2-b546-9344f0cb72a7") : secret "infra-operator-webhook-server-cert" not found Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.665229 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6ndx\" (UniqueName: \"kubernetes.io/projected/10978ce4-0034-4ad4-9a47-2f1931c09655-kube-api-access-r6ndx\") pod \"keystone-operator-controller-manager-ddb98f99b-h2bt5\" (UID: \"10978ce4-0034-4ad4-9a47-2f1931c09655\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-h2bt5" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.667210 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-xxvtr"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.668240 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h256t\" (UniqueName: \"kubernetes.io/projected/9fdf7ed4-1e18-43c0-b305-2b9b226f4634-kube-api-access-h256t\") pod \"ironic-operator-controller-manager-74cb5cbc49-d5qnn\" (UID: \"9fdf7ed4-1e18-43c0-b305-2b9b226f4634\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-d5qnn" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.668267 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2cx6\" (UniqueName: \"kubernetes.io/projected/fc401fc1-e03b-4b32-8782-15e9c2a70e8a-kube-api-access-v2cx6\") pod \"manila-operator-controller-manager-59578bc799-t9wqw\" (UID: \"fc401fc1-e03b-4b32-8782-15e9c2a70e8a\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-t9wqw" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.668315 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-xxvtr" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.672767 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjbw9\" (UniqueName: \"kubernetes.io/projected/443416f7-e4ce-41f2-b546-9344f0cb72a7-kube-api-access-kjbw9\") pod \"infra-operator-controller-manager-585fc5b659-dt9kx\" (UID: \"443416f7-e4ce-41f2-b546-9344f0cb72a7\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.673541 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-6bbfn" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.683557 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-xxvtr"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.684656 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-t9wqw" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.699108 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-qmbvh" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.739027 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.740148 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.742857 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.747161 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-nzjx6" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.754172 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-rz48f"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.755127 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-rz48f" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.755236 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b628\" (UniqueName: \"kubernetes.io/projected/83b0fe55-090f-448e-95c2-1d3cbbc1b54c-kube-api-access-9b628\") pod \"neutron-operator-controller-manager-797d478b46-2dmpc\" (UID: \"83b0fe55-090f-448e-95c2-1d3cbbc1b54c\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2dmpc" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.755266 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhngf\" (UniqueName: \"kubernetes.io/projected/420fdd11-0131-479d-9f7a-47f091cacd83-kube-api-access-nhngf\") pod \"mariadb-operator-controller-manager-5777b4f897-vc9rk\" (UID: \"420fdd11-0131-479d-9f7a-47f091cacd83\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-vc9rk" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.755291 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bkrg\" (UniqueName: \"kubernetes.io/projected/93defbbc-c732-4743-bcff-a1d7f06aefa2-kube-api-access-2bkrg\") pod \"nova-operator-controller-manager-57bb74c7bf-89lzv\" (UID: \"93defbbc-c732-4743-bcff-a1d7f06aefa2\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-89lzv" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.755316 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n96b\" (UniqueName: \"kubernetes.io/projected/e1c2fd77-699a-4c79-9c54-9251c5ad7d17-kube-api-access-8n96b\") pod \"octavia-operator-controller-manager-6d7c7ddf95-xxvtr\" (UID: \"e1c2fd77-699a-4c79-9c54-9251c5ad7d17\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-xxvtr" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.757667 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-jvqg4" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.768833 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.772537 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-b7w5k"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.776476 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-b7w5k" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.779215 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-nz9xg" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.781568 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b628\" (UniqueName: \"kubernetes.io/projected/83b0fe55-090f-448e-95c2-1d3cbbc1b54c-kube-api-access-9b628\") pod \"neutron-operator-controller-manager-797d478b46-2dmpc\" (UID: \"83b0fe55-090f-448e-95c2-1d3cbbc1b54c\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2dmpc" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.783573 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhngf\" (UniqueName: \"kubernetes.io/projected/420fdd11-0131-479d-9f7a-47f091cacd83-kube-api-access-nhngf\") pod \"mariadb-operator-controller-manager-5777b4f897-vc9rk\" (UID: \"420fdd11-0131-479d-9f7a-47f091cacd83\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-vc9rk" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.785688 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-rz48f"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.797592 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qzh4v"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.798891 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qzh4v" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.806976 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-b7w5k"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.809526 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qzh4v"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.817934 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-d5qnn" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.818880 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-b55bc" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.836020 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-h4jn2"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.838975 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-h4jn2" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.843909 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-h4jn2"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.845232 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-h2bt5" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.852625 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-r4w9n" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.856915 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sjf9\" (UniqueName: \"kubernetes.io/projected/0ec7257b-b487-453c-88e8-d279184acdfc-kube-api-access-6sjf9\") pod \"ovn-operator-controller-manager-869cc7797f-rz48f\" (UID: \"0ec7257b-b487-453c-88e8-d279184acdfc\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-rz48f" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.856967 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n96b\" (UniqueName: \"kubernetes.io/projected/e1c2fd77-699a-4c79-9c54-9251c5ad7d17-kube-api-access-8n96b\") pod \"octavia-operator-controller-manager-6d7c7ddf95-xxvtr\" (UID: \"e1c2fd77-699a-4c79-9c54-9251c5ad7d17\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-xxvtr" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.857228 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv8lh\" (UniqueName: \"kubernetes.io/projected/36db8827-b9af-4b9b-87aa-3cfa7d005ee0-kube-api-access-hv8lh\") pod \"placement-operator-controller-manager-664664cb68-b7w5k\" (UID: \"36db8827-b9af-4b9b-87aa-3cfa7d005ee0\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-b7w5k" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.857280 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd5tw\" (UniqueName: \"kubernetes.io/projected/0174cdbc-06e0-423f-af63-5467737e6ec5-kube-api-access-jd5tw\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw\" (UID: \"0174cdbc-06e0-423f-af63-5467737e6ec5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.857306 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khqb8\" (UniqueName: \"kubernetes.io/projected/e6828db3-0a2c-4b98-8cbf-1eba04f13bc5-kube-api-access-khqb8\") pod \"swift-operator-controller-manager-5f4d5dfdc6-qzh4v\" (UID: \"e6828db3-0a2c-4b98-8cbf-1eba04f13bc5\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qzh4v" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.857353 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0174cdbc-06e0-423f-af63-5467737e6ec5-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw\" (UID: \"0174cdbc-06e0-423f-af63-5467737e6ec5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.857771 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bkrg\" (UniqueName: \"kubernetes.io/projected/93defbbc-c732-4743-bcff-a1d7f06aefa2-kube-api-access-2bkrg\") pod \"nova-operator-controller-manager-57bb74c7bf-89lzv\" (UID: \"93defbbc-c732-4743-bcff-a1d7f06aefa2\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-89lzv" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.878535 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n96b\" (UniqueName: \"kubernetes.io/projected/e1c2fd77-699a-4c79-9c54-9251c5ad7d17-kube-api-access-8n96b\") pod \"octavia-operator-controller-manager-6d7c7ddf95-xxvtr\" (UID: \"e1c2fd77-699a-4c79-9c54-9251c5ad7d17\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-xxvtr" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.879528 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bkrg\" (UniqueName: \"kubernetes.io/projected/93defbbc-c732-4743-bcff-a1d7f06aefa2-kube-api-access-2bkrg\") pod \"nova-operator-controller-manager-57bb74c7bf-89lzv\" (UID: \"93defbbc-c732-4743-bcff-a1d7f06aefa2\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-89lzv" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.882092 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-lr987"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.885747 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-lr987" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.887860 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-lr987"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.891590 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-bntw9" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.920965 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-96mn7"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.922244 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-646675d848-96mn7" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.928933 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-96mn7"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.932626 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-ch9jz" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.949449 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc"] Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.950562 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.954924 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.955152 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-sfvcg" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.959450 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2wsz\" (UniqueName: \"kubernetes.io/projected/ba76d6db-10cc-4d7d-b620-e37533730395-kube-api-access-t2wsz\") pod \"test-operator-controller-manager-ffcdd6c94-lr987\" (UID: \"ba76d6db-10cc-4d7d-b620-e37533730395\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-lr987" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.959514 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv8lh\" (UniqueName: \"kubernetes.io/projected/36db8827-b9af-4b9b-87aa-3cfa7d005ee0-kube-api-access-hv8lh\") pod \"placement-operator-controller-manager-664664cb68-b7w5k\" (UID: \"36db8827-b9af-4b9b-87aa-3cfa7d005ee0\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-b7w5k" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.959545 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd5tw\" (UniqueName: \"kubernetes.io/projected/0174cdbc-06e0-423f-af63-5467737e6ec5-kube-api-access-jd5tw\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw\" (UID: \"0174cdbc-06e0-423f-af63-5467737e6ec5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.959567 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khqb8\" (UniqueName: \"kubernetes.io/projected/e6828db3-0a2c-4b98-8cbf-1eba04f13bc5-kube-api-access-khqb8\") pod \"swift-operator-controller-manager-5f4d5dfdc6-qzh4v\" (UID: \"e6828db3-0a2c-4b98-8cbf-1eba04f13bc5\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qzh4v" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.959595 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0174cdbc-06e0-423f-af63-5467737e6ec5-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw\" (UID: \"0174cdbc-06e0-423f-af63-5467737e6ec5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.959619 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sjf9\" (UniqueName: \"kubernetes.io/projected/0ec7257b-b487-453c-88e8-d279184acdfc-kube-api-access-6sjf9\") pod \"ovn-operator-controller-manager-869cc7797f-rz48f\" (UID: \"0ec7257b-b487-453c-88e8-d279184acdfc\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-rz48f" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.959655 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jdxx\" (UniqueName: \"kubernetes.io/projected/a5094dca-dde7-4144-bfbc-b4f90c170a21-kube-api-access-5jdxx\") pod \"telemetry-operator-controller-manager-578874c84d-h4jn2\" (UID: \"a5094dca-dde7-4144-bfbc-b4f90c170a21\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-h4jn2" Oct 10 13:30:53 crc kubenswrapper[4745]: E1010 13:30:53.959828 4745 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 10 13:30:53 crc kubenswrapper[4745]: E1010 13:30:53.959882 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0174cdbc-06e0-423f-af63-5467737e6ec5-cert podName:0174cdbc-06e0-423f-af63-5467737e6ec5 nodeName:}" failed. No retries permitted until 2025-10-10 13:30:54.459866245 +0000 UTC m=+768.357523008 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0174cdbc-06e0-423f-af63-5467737e6ec5-cert") pod "openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw" (UID: "0174cdbc-06e0-423f-af63-5467737e6ec5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.990071 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv8lh\" (UniqueName: \"kubernetes.io/projected/36db8827-b9af-4b9b-87aa-3cfa7d005ee0-kube-api-access-hv8lh\") pod \"placement-operator-controller-manager-664664cb68-b7w5k\" (UID: \"36db8827-b9af-4b9b-87aa-3cfa7d005ee0\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-b7w5k" Oct 10 13:30:53 crc kubenswrapper[4745]: I1010 13:30:53.991263 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc"] Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:53.999134 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-vc9rk" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.003379 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sjf9\" (UniqueName: \"kubernetes.io/projected/0ec7257b-b487-453c-88e8-d279184acdfc-kube-api-access-6sjf9\") pod \"ovn-operator-controller-manager-869cc7797f-rz48f\" (UID: \"0ec7257b-b487-453c-88e8-d279184acdfc\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-rz48f" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.004894 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khqb8\" (UniqueName: \"kubernetes.io/projected/e6828db3-0a2c-4b98-8cbf-1eba04f13bc5-kube-api-access-khqb8\") pod \"swift-operator-controller-manager-5f4d5dfdc6-qzh4v\" (UID: \"e6828db3-0a2c-4b98-8cbf-1eba04f13bc5\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qzh4v" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.014159 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd5tw\" (UniqueName: \"kubernetes.io/projected/0174cdbc-06e0-423f-af63-5467737e6ec5-kube-api-access-jd5tw\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw\" (UID: \"0174cdbc-06e0-423f-af63-5467737e6ec5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.030524 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2dmpc" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.043357 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-89lzv" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.047599 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c9gmp"] Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.048598 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c9gmp" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.051277 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-dfz8z" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.053032 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-xxvtr" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.053485 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c9gmp"] Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.065022 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jdxx\" (UniqueName: \"kubernetes.io/projected/a5094dca-dde7-4144-bfbc-b4f90c170a21-kube-api-access-5jdxx\") pod \"telemetry-operator-controller-manager-578874c84d-h4jn2\" (UID: \"a5094dca-dde7-4144-bfbc-b4f90c170a21\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-h4jn2" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.065085 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2p2v\" (UniqueName: \"kubernetes.io/projected/fd4f5298-2862-4650-950d-abd7f8195942-kube-api-access-r2p2v\") pod \"openstack-operator-controller-manager-6f95c9794b-jtlvc\" (UID: \"fd4f5298-2862-4650-950d-abd7f8195942\") " pod="openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.065113 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2wsz\" (UniqueName: \"kubernetes.io/projected/ba76d6db-10cc-4d7d-b620-e37533730395-kube-api-access-t2wsz\") pod \"test-operator-controller-manager-ffcdd6c94-lr987\" (UID: \"ba76d6db-10cc-4d7d-b620-e37533730395\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-lr987" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.065137 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd4f5298-2862-4650-950d-abd7f8195942-cert\") pod \"openstack-operator-controller-manager-6f95c9794b-jtlvc\" (UID: \"fd4f5298-2862-4650-950d-abd7f8195942\") " pod="openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.065155 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mphr\" (UniqueName: \"kubernetes.io/projected/6d02fe8c-f0b4-4ef1-8e57-569a836708cd-kube-api-access-4mphr\") pod \"watcher-operator-controller-manager-646675d848-96mn7\" (UID: \"6d02fe8c-f0b4-4ef1-8e57-569a836708cd\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-96mn7" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.091196 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2wsz\" (UniqueName: \"kubernetes.io/projected/ba76d6db-10cc-4d7d-b620-e37533730395-kube-api-access-t2wsz\") pod \"test-operator-controller-manager-ffcdd6c94-lr987\" (UID: \"ba76d6db-10cc-4d7d-b620-e37533730395\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-lr987" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.094456 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jdxx\" (UniqueName: \"kubernetes.io/projected/a5094dca-dde7-4144-bfbc-b4f90c170a21-kube-api-access-5jdxx\") pod \"telemetry-operator-controller-manager-578874c84d-h4jn2\" (UID: \"a5094dca-dde7-4144-bfbc-b4f90c170a21\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-h4jn2" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.166797 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd4f5298-2862-4650-950d-abd7f8195942-cert\") pod \"openstack-operator-controller-manager-6f95c9794b-jtlvc\" (UID: \"fd4f5298-2862-4650-950d-abd7f8195942\") " pod="openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.166838 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mphr\" (UniqueName: \"kubernetes.io/projected/6d02fe8c-f0b4-4ef1-8e57-569a836708cd-kube-api-access-4mphr\") pod \"watcher-operator-controller-manager-646675d848-96mn7\" (UID: \"6d02fe8c-f0b4-4ef1-8e57-569a836708cd\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-96mn7" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.166865 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2jgs\" (UniqueName: \"kubernetes.io/projected/ebc60df2-d5b0-418f-a8cd-e8440706d3be-kube-api-access-f2jgs\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-c9gmp\" (UID: \"ebc60df2-d5b0-418f-a8cd-e8440706d3be\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c9gmp" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.166941 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/443416f7-e4ce-41f2-b546-9344f0cb72a7-cert\") pod \"infra-operator-controller-manager-585fc5b659-dt9kx\" (UID: \"443416f7-e4ce-41f2-b546-9344f0cb72a7\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx" Oct 10 13:30:54 crc kubenswrapper[4745]: E1010 13:30:54.166959 4745 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.166972 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2p2v\" (UniqueName: \"kubernetes.io/projected/fd4f5298-2862-4650-950d-abd7f8195942-kube-api-access-r2p2v\") pod \"openstack-operator-controller-manager-6f95c9794b-jtlvc\" (UID: \"fd4f5298-2862-4650-950d-abd7f8195942\") " pod="openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc" Oct 10 13:30:54 crc kubenswrapper[4745]: E1010 13:30:54.167144 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fd4f5298-2862-4650-950d-abd7f8195942-cert podName:fd4f5298-2862-4650-950d-abd7f8195942 nodeName:}" failed. No retries permitted until 2025-10-10 13:30:54.667122898 +0000 UTC m=+768.564779661 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fd4f5298-2862-4650-950d-abd7f8195942-cert") pod "openstack-operator-controller-manager-6f95c9794b-jtlvc" (UID: "fd4f5298-2862-4650-950d-abd7f8195942") : secret "webhook-server-cert" not found Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.170411 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/443416f7-e4ce-41f2-b546-9344f0cb72a7-cert\") pod \"infra-operator-controller-manager-585fc5b659-dt9kx\" (UID: \"443416f7-e4ce-41f2-b546-9344f0cb72a7\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.185143 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2p2v\" (UniqueName: \"kubernetes.io/projected/fd4f5298-2862-4650-950d-abd7f8195942-kube-api-access-r2p2v\") pod \"openstack-operator-controller-manager-6f95c9794b-jtlvc\" (UID: \"fd4f5298-2862-4650-950d-abd7f8195942\") " pod="openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.190832 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mphr\" (UniqueName: \"kubernetes.io/projected/6d02fe8c-f0b4-4ef1-8e57-569a836708cd-kube-api-access-4mphr\") pod \"watcher-operator-controller-manager-646675d848-96mn7\" (UID: \"6d02fe8c-f0b4-4ef1-8e57-569a836708cd\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-96mn7" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.192274 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-rz48f" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.223618 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-b7w5k" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.256218 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qzh4v" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.267702 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-htmcv"] Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.268510 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2jgs\" (UniqueName: \"kubernetes.io/projected/ebc60df2-d5b0-418f-a8cd-e8440706d3be-kube-api-access-f2jgs\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-c9gmp\" (UID: \"ebc60df2-d5b0-418f-a8cd-e8440706d3be\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c9gmp" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.288162 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2jgs\" (UniqueName: \"kubernetes.io/projected/ebc60df2-d5b0-418f-a8cd-e8440706d3be-kube-api-access-f2jgs\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-c9gmp\" (UID: \"ebc60df2-d5b0-418f-a8cd-e8440706d3be\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c9gmp" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.306055 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-h4jn2" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.333651 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-lr987" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.358417 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.391721 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-646675d848-96mn7" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.426037 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7g9x6" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.453380 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c9gmp" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.472760 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0174cdbc-06e0-423f-af63-5467737e6ec5-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw\" (UID: \"0174cdbc-06e0-423f-af63-5467737e6ec5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.479291 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0174cdbc-06e0-423f-af63-5467737e6ec5-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw\" (UID: \"0174cdbc-06e0-423f-af63-5467737e6ec5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.521869 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-htmcv" event={"ID":"730b5555-5f6d-4e16-91e0-faacad56d777","Type":"ContainerStarted","Data":"a59d43af5d97fd93f161c734c51e2c3befcfc58dc4f57a3acbc25d4152c302be"} Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.524422 4745 generic.go:334] "Generic (PLEG): container finished" podID="2dd7d5df-eee1-4f07-8dd8-24873ec41827" containerID="4c35453f0d240185f0b98fafd3e9fde2ea6ed29883dfa2109c3f746a1964afda" exitCode=0 Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.524448 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7g9x6" event={"ID":"2dd7d5df-eee1-4f07-8dd8-24873ec41827","Type":"ContainerDied","Data":"4c35453f0d240185f0b98fafd3e9fde2ea6ed29883dfa2109c3f746a1964afda"} Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.524465 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7g9x6" event={"ID":"2dd7d5df-eee1-4f07-8dd8-24873ec41827","Type":"ContainerDied","Data":"bbeefee00cb77bdf6f9cf1a8cdf816c680a63f5802ccb0d12067e6cb628ae1ff"} Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.524483 4745 scope.go:117] "RemoveContainer" containerID="4c35453f0d240185f0b98fafd3e9fde2ea6ed29883dfa2109c3f746a1964afda" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.524583 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7g9x6" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.573775 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd7d5df-eee1-4f07-8dd8-24873ec41827-catalog-content\") pod \"2dd7d5df-eee1-4f07-8dd8-24873ec41827\" (UID: \"2dd7d5df-eee1-4f07-8dd8-24873ec41827\") " Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.573926 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gnvh\" (UniqueName: \"kubernetes.io/projected/2dd7d5df-eee1-4f07-8dd8-24873ec41827-kube-api-access-6gnvh\") pod \"2dd7d5df-eee1-4f07-8dd8-24873ec41827\" (UID: \"2dd7d5df-eee1-4f07-8dd8-24873ec41827\") " Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.574001 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd7d5df-eee1-4f07-8dd8-24873ec41827-utilities\") pod \"2dd7d5df-eee1-4f07-8dd8-24873ec41827\" (UID: \"2dd7d5df-eee1-4f07-8dd8-24873ec41827\") " Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.575262 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dd7d5df-eee1-4f07-8dd8-24873ec41827-utilities" (OuterVolumeSpecName: "utilities") pod "2dd7d5df-eee1-4f07-8dd8-24873ec41827" (UID: "2dd7d5df-eee1-4f07-8dd8-24873ec41827"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.584929 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dd7d5df-eee1-4f07-8dd8-24873ec41827-kube-api-access-6gnvh" (OuterVolumeSpecName: "kube-api-access-6gnvh") pod "2dd7d5df-eee1-4f07-8dd8-24873ec41827" (UID: "2dd7d5df-eee1-4f07-8dd8-24873ec41827"). InnerVolumeSpecName "kube-api-access-6gnvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.589363 4745 scope.go:117] "RemoveContainer" containerID="52d5539698174948775e905fdf84bbc76aff1b853796e85220df5bb3264ee9ba" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.624468 4745 scope.go:117] "RemoveContainer" containerID="34df699941d4a668bc2b247cec10405629512ec51f65fdf8573b5513769d99cd" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.644557 4745 scope.go:117] "RemoveContainer" containerID="4c35453f0d240185f0b98fafd3e9fde2ea6ed29883dfa2109c3f746a1964afda" Oct 10 13:30:54 crc kubenswrapper[4745]: E1010 13:30:54.645120 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c35453f0d240185f0b98fafd3e9fde2ea6ed29883dfa2109c3f746a1964afda\": container with ID starting with 4c35453f0d240185f0b98fafd3e9fde2ea6ed29883dfa2109c3f746a1964afda not found: ID does not exist" containerID="4c35453f0d240185f0b98fafd3e9fde2ea6ed29883dfa2109c3f746a1964afda" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.645154 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c35453f0d240185f0b98fafd3e9fde2ea6ed29883dfa2109c3f746a1964afda"} err="failed to get container status \"4c35453f0d240185f0b98fafd3e9fde2ea6ed29883dfa2109c3f746a1964afda\": rpc error: code = NotFound desc = could not find container \"4c35453f0d240185f0b98fafd3e9fde2ea6ed29883dfa2109c3f746a1964afda\": container with ID starting with 4c35453f0d240185f0b98fafd3e9fde2ea6ed29883dfa2109c3f746a1964afda not found: ID does not exist" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.645180 4745 scope.go:117] "RemoveContainer" containerID="52d5539698174948775e905fdf84bbc76aff1b853796e85220df5bb3264ee9ba" Oct 10 13:30:54 crc kubenswrapper[4745]: E1010 13:30:54.645800 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52d5539698174948775e905fdf84bbc76aff1b853796e85220df5bb3264ee9ba\": container with ID starting with 52d5539698174948775e905fdf84bbc76aff1b853796e85220df5bb3264ee9ba not found: ID does not exist" containerID="52d5539698174948775e905fdf84bbc76aff1b853796e85220df5bb3264ee9ba" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.645838 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52d5539698174948775e905fdf84bbc76aff1b853796e85220df5bb3264ee9ba"} err="failed to get container status \"52d5539698174948775e905fdf84bbc76aff1b853796e85220df5bb3264ee9ba\": rpc error: code = NotFound desc = could not find container \"52d5539698174948775e905fdf84bbc76aff1b853796e85220df5bb3264ee9ba\": container with ID starting with 52d5539698174948775e905fdf84bbc76aff1b853796e85220df5bb3264ee9ba not found: ID does not exist" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.645867 4745 scope.go:117] "RemoveContainer" containerID="34df699941d4a668bc2b247cec10405629512ec51f65fdf8573b5513769d99cd" Oct 10 13:30:54 crc kubenswrapper[4745]: E1010 13:30:54.646200 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34df699941d4a668bc2b247cec10405629512ec51f65fdf8573b5513769d99cd\": container with ID starting with 34df699941d4a668bc2b247cec10405629512ec51f65fdf8573b5513769d99cd not found: ID does not exist" containerID="34df699941d4a668bc2b247cec10405629512ec51f65fdf8573b5513769d99cd" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.646223 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34df699941d4a668bc2b247cec10405629512ec51f65fdf8573b5513769d99cd"} err="failed to get container status \"34df699941d4a668bc2b247cec10405629512ec51f65fdf8573b5513769d99cd\": rpc error: code = NotFound desc = could not find container \"34df699941d4a668bc2b247cec10405629512ec51f65fdf8573b5513769d99cd\": container with ID starting with 34df699941d4a668bc2b247cec10405629512ec51f65fdf8573b5513769d99cd not found: ID does not exist" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.676483 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd4f5298-2862-4650-950d-abd7f8195942-cert\") pod \"openstack-operator-controller-manager-6f95c9794b-jtlvc\" (UID: \"fd4f5298-2862-4650-950d-abd7f8195942\") " pod="openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.676562 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd7d5df-eee1-4f07-8dd8-24873ec41827-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.676581 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gnvh\" (UniqueName: \"kubernetes.io/projected/2dd7d5df-eee1-4f07-8dd8-24873ec41827-kube-api-access-6gnvh\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:54 crc kubenswrapper[4745]: E1010 13:30:54.676699 4745 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 10 13:30:54 crc kubenswrapper[4745]: E1010 13:30:54.676782 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fd4f5298-2862-4650-950d-abd7f8195942-cert podName:fd4f5298-2862-4650-950d-abd7f8195942 nodeName:}" failed. No retries permitted until 2025-10-10 13:30:55.676764379 +0000 UTC m=+769.574421142 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fd4f5298-2862-4650-950d-abd7f8195942-cert") pod "openstack-operator-controller-manager-6f95c9794b-jtlvc" (UID: "fd4f5298-2862-4650-950d-abd7f8195942") : secret "webhook-server-cert" not found Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.680834 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.681599 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dd7d5df-eee1-4f07-8dd8-24873ec41827-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2dd7d5df-eee1-4f07-8dd8-24873ec41827" (UID: "2dd7d5df-eee1-4f07-8dd8-24873ec41827"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.778522 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd7d5df-eee1-4f07-8dd8-24873ec41827-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.816287 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tpb4j"] Oct 10 13:30:54 crc kubenswrapper[4745]: E1010 13:30:54.816679 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dd7d5df-eee1-4f07-8dd8-24873ec41827" containerName="extract-utilities" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.816707 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dd7d5df-eee1-4f07-8dd8-24873ec41827" containerName="extract-utilities" Oct 10 13:30:54 crc kubenswrapper[4745]: E1010 13:30:54.816721 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dd7d5df-eee1-4f07-8dd8-24873ec41827" containerName="extract-content" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.816741 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dd7d5df-eee1-4f07-8dd8-24873ec41827" containerName="extract-content" Oct 10 13:30:54 crc kubenswrapper[4745]: E1010 13:30:54.816751 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dd7d5df-eee1-4f07-8dd8-24873ec41827" containerName="registry-server" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.816759 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dd7d5df-eee1-4f07-8dd8-24873ec41827" containerName="registry-server" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.816935 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dd7d5df-eee1-4f07-8dd8-24873ec41827" containerName="registry-server" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.818089 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tpb4j" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.822682 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tpb4j"] Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.849521 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7g9x6"] Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.861905 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7g9x6"] Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.880049 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp545\" (UniqueName: \"kubernetes.io/projected/15e38293-23c6-4705-9aaf-1de10eb3fe1e-kube-api-access-cp545\") pod \"certified-operators-tpb4j\" (UID: \"15e38293-23c6-4705-9aaf-1de10eb3fe1e\") " pod="openshift-marketplace/certified-operators-tpb4j" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.880149 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15e38293-23c6-4705-9aaf-1de10eb3fe1e-utilities\") pod \"certified-operators-tpb4j\" (UID: \"15e38293-23c6-4705-9aaf-1de10eb3fe1e\") " pod="openshift-marketplace/certified-operators-tpb4j" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.880199 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15e38293-23c6-4705-9aaf-1de10eb3fe1e-catalog-content\") pod \"certified-operators-tpb4j\" (UID: \"15e38293-23c6-4705-9aaf-1de10eb3fe1e\") " pod="openshift-marketplace/certified-operators-tpb4j" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.906721 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-2w6nw"] Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.914800 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-fpcxk"] Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.936249 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-t9wqw"] Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.950823 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-d5qnn"] Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.962371 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-tz665"] Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.969449 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-h2bt5"] Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.981468 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15e38293-23c6-4705-9aaf-1de10eb3fe1e-utilities\") pod \"certified-operators-tpb4j\" (UID: \"15e38293-23c6-4705-9aaf-1de10eb3fe1e\") " pod="openshift-marketplace/certified-operators-tpb4j" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.981656 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15e38293-23c6-4705-9aaf-1de10eb3fe1e-catalog-content\") pod \"certified-operators-tpb4j\" (UID: \"15e38293-23c6-4705-9aaf-1de10eb3fe1e\") " pod="openshift-marketplace/certified-operators-tpb4j" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.981896 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp545\" (UniqueName: \"kubernetes.io/projected/15e38293-23c6-4705-9aaf-1de10eb3fe1e-kube-api-access-cp545\") pod \"certified-operators-tpb4j\" (UID: \"15e38293-23c6-4705-9aaf-1de10eb3fe1e\") " pod="openshift-marketplace/certified-operators-tpb4j" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.981973 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15e38293-23c6-4705-9aaf-1de10eb3fe1e-utilities\") pod \"certified-operators-tpb4j\" (UID: \"15e38293-23c6-4705-9aaf-1de10eb3fe1e\") " pod="openshift-marketplace/certified-operators-tpb4j" Oct 10 13:30:54 crc kubenswrapper[4745]: W1010 13:30:54.983221 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ca519af_249e_48f0_8669_77bd548be8f3.slice/crio-d96128c22945f6d4ebd35346ae5116adcb79faa2759584002790d3df3db32e6b WatchSource:0}: Error finding container d96128c22945f6d4ebd35346ae5116adcb79faa2759584002790d3df3db32e6b: Status 404 returned error can't find the container with id d96128c22945f6d4ebd35346ae5116adcb79faa2759584002790d3df3db32e6b Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.983490 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15e38293-23c6-4705-9aaf-1de10eb3fe1e-catalog-content\") pod \"certified-operators-tpb4j\" (UID: \"15e38293-23c6-4705-9aaf-1de10eb3fe1e\") " pod="openshift-marketplace/certified-operators-tpb4j" Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.984062 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-c7986"] Oct 10 13:30:54 crc kubenswrapper[4745]: W1010 13:30:54.995890 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc401fc1_e03b_4b32_8782_15e9c2a70e8a.slice/crio-b24d0649cc881e61bdff30a7aa0f9b6959703809d0992a0cf34cee1e3fed96ba WatchSource:0}: Error finding container b24d0649cc881e61bdff30a7aa0f9b6959703809d0992a0cf34cee1e3fed96ba: Status 404 returned error can't find the container with id b24d0649cc881e61bdff30a7aa0f9b6959703809d0992a0cf34cee1e3fed96ba Oct 10 13:30:54 crc kubenswrapper[4745]: I1010 13:30:54.997976 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp545\" (UniqueName: \"kubernetes.io/projected/15e38293-23c6-4705-9aaf-1de10eb3fe1e-kube-api-access-cp545\") pod \"certified-operators-tpb4j\" (UID: \"15e38293-23c6-4705-9aaf-1de10eb3fe1e\") " pod="openshift-marketplace/certified-operators-tpb4j" Oct 10 13:30:54 crc kubenswrapper[4745]: W1010 13:30:54.999211 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddff847b8_915b_4b64_85c3_d7c20d1282a0.slice/crio-e4805b5e47161efeafa23b51d945188d432ebcda974a784f9a00df01e517633f WatchSource:0}: Error finding container e4805b5e47161efeafa23b51d945188d432ebcda974a784f9a00df01e517633f: Status 404 returned error can't find the container with id e4805b5e47161efeafa23b51d945188d432ebcda974a784f9a00df01e517633f Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.118364 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-qmbvh"] Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.118414 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-89lzv"] Oct 10 13:30:55 crc kubenswrapper[4745]: W1010 13:30:55.123574 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93defbbc_c732_4743_bcff_a1d7f06aefa2.slice/crio-0354828870be0d5c51285bb696897358d4501c6cd280b15a45deb2b8f679c870 WatchSource:0}: Error finding container 0354828870be0d5c51285bb696897358d4501c6cd280b15a45deb2b8f679c870: Status 404 returned error can't find the container with id 0354828870be0d5c51285bb696897358d4501c6cd280b15a45deb2b8f679c870 Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.135024 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-vc9rk"] Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.135168 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tpb4j" Oct 10 13:30:55 crc kubenswrapper[4745]: W1010 13:30:55.173847 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod420fdd11_0131_479d_9f7a_47f091cacd83.slice/crio-8c3149ee75ab54c230c674afdb068dc3558c520b9134f2082f8b471a2b9010f6 WatchSource:0}: Error finding container 8c3149ee75ab54c230c674afdb068dc3558c520b9134f2082f8b471a2b9010f6: Status 404 returned error can't find the container with id 8c3149ee75ab54c230c674afdb068dc3558c520b9134f2082f8b471a2b9010f6 Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.498672 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c9gmp"] Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.520120 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-lr987"] Oct 10 13:30:55 crc kubenswrapper[4745]: W1010 13:30:55.534624 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5094dca_dde7_4144_bfbc_b4f90c170a21.slice/crio-fb7f5ab48460b5d3d9eb38315b5d6f298131069dfa174cc8a232e0e4f0d01970 WatchSource:0}: Error finding container fb7f5ab48460b5d3d9eb38315b5d6f298131069dfa174cc8a232e0e4f0d01970: Status 404 returned error can't find the container with id fb7f5ab48460b5d3d9eb38315b5d6f298131069dfa174cc8a232e0e4f0d01970 Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.534628 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-h4jn2"] Oct 10 13:30:55 crc kubenswrapper[4745]: W1010 13:30:55.535476 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba76d6db_10cc_4d7d_b620_e37533730395.slice/crio-5acf01f624ebf23fcab4992b34eb6ea58a6e2d218b19beeeccad320c3195986a WatchSource:0}: Error finding container 5acf01f624ebf23fcab4992b34eb6ea58a6e2d218b19beeeccad320c3195986a: Status 404 returned error can't find the container with id 5acf01f624ebf23fcab4992b34eb6ea58a6e2d218b19beeeccad320c3195986a Oct 10 13:30:55 crc kubenswrapper[4745]: W1010 13:30:55.538794 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod443416f7_e4ce_41f2_b546_9344f0cb72a7.slice/crio-48ec6bc9f2f512fd544daaecfa0e509a9bc4fd29b8d858f1fd6cdac8617b121f WatchSource:0}: Error finding container 48ec6bc9f2f512fd544daaecfa0e509a9bc4fd29b8d858f1fd6cdac8617b121f: Status 404 returned error can't find the container with id 48ec6bc9f2f512fd544daaecfa0e509a9bc4fd29b8d858f1fd6cdac8617b121f Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.551663 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-xxvtr"] Oct 10 13:30:55 crc kubenswrapper[4745]: E1010 13:30:55.561290 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t2wsz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-ffcdd6c94-lr987_openstack-operators(ba76d6db-10cc-4d7d-b620-e37533730395): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.564839 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-c7986" event={"ID":"5ca519af-249e-48f0-8669-77bd548be8f3","Type":"ContainerStarted","Data":"d96128c22945f6d4ebd35346ae5116adcb79faa2759584002790d3df3db32e6b"} Oct 10 13:30:55 crc kubenswrapper[4745]: E1010 13:30:55.567672 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9b628,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-797d478b46-2dmpc_openstack-operators(83b0fe55-090f-448e-95c2-1d3cbbc1b54c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.571478 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-96mn7"] Oct 10 13:30:55 crc kubenswrapper[4745]: E1010 13:30:55.573259 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6sjf9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-869cc7797f-rz48f_openstack-operators(0ec7257b-b487-453c-88e8-d279184acdfc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.579342 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-d5qnn" event={"ID":"9fdf7ed4-1e18-43c0-b305-2b9b226f4634","Type":"ContainerStarted","Data":"42a2d39a5fb5877ffcc7ca8e0e9910382e5b493b3180363e433936cafa1d87aa"} Oct 10 13:30:55 crc kubenswrapper[4745]: E1010 13:30:55.585592 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hv8lh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-664664cb68-b7w5k_openstack-operators(36db8827-b9af-4b9b-87aa-3cfa7d005ee0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.585719 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-vc9rk" event={"ID":"420fdd11-0131-479d-9f7a-47f091cacd83","Type":"ContainerStarted","Data":"8c3149ee75ab54c230c674afdb068dc3558c520b9134f2082f8b471a2b9010f6"} Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.591002 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-t9wqw" event={"ID":"fc401fc1-e03b-4b32-8782-15e9c2a70e8a","Type":"ContainerStarted","Data":"b24d0649cc881e61bdff30a7aa0f9b6959703809d0992a0cf34cee1e3fed96ba"} Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.607970 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx"] Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.623781 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-rz48f"] Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.630365 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-2dmpc"] Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.631751 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-b7w5k"] Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.633548 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tz665" event={"ID":"446e0b63-81e5-47fb-b110-cfa5393e5311","Type":"ContainerStarted","Data":"97a799f4402f69b47ca49dbf5e6ad0da6b41f5722c47bdbd701a7f44aec0cbb4"} Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.635303 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tpb4j"] Oct 10 13:30:55 crc kubenswrapper[4745]: E1010 13:30:55.658188 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4mphr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-646675d848-96mn7_openstack-operators(6d02fe8c-f0b4-4ef1-8e57-569a836708cd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 10 13:30:55 crc kubenswrapper[4745]: E1010 13:30:55.658286 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:09deecf840d38ff6af3c924729cf0a9444bc985848bfbe7c918019b88a6bc4d7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8n96b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-6d7c7ddf95-xxvtr_openstack-operators(e1c2fd77-699a-4c79-9c54-9251c5ad7d17): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.661545 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-qmbvh" event={"ID":"1d2b124e-5495-42bd-b675-4eed70ccc446","Type":"ContainerStarted","Data":"3f0367d2fcd895d9f19dc194c28cbfa1065555e68f4bdfa3cc3a3371df6ba4b3"} Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.674954 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qzh4v"] Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.688958 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw"] Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.690610 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-2w6nw" event={"ID":"7f07678c-0cc2-47a5-b244-72ae3aa55baa","Type":"ContainerStarted","Data":"ed33c9d218a287bee6b4403a49f96ffad1921e474810b7a3cda504213f101619"} Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.691869 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-h2bt5" event={"ID":"10978ce4-0034-4ad4-9a47-2f1931c09655","Type":"ContainerStarted","Data":"378b36ada2241d2a40f2c541a16f127b75a804eba6509fcc8abe47cf98fa9467"} Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.693272 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-fpcxk" event={"ID":"dff847b8-915b-4b64-85c3-d7c20d1282a0","Type":"ContainerStarted","Data":"e4805b5e47161efeafa23b51d945188d432ebcda974a784f9a00df01e517633f"} Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.694311 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-89lzv" event={"ID":"93defbbc-c732-4743-bcff-a1d7f06aefa2","Type":"ContainerStarted","Data":"0354828870be0d5c51285bb696897358d4501c6cd280b15a45deb2b8f679c870"} Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.716854 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd4f5298-2862-4650-950d-abd7f8195942-cert\") pod \"openstack-operator-controller-manager-6f95c9794b-jtlvc\" (UID: \"fd4f5298-2862-4650-950d-abd7f8195942\") " pod="openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc" Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.724468 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd4f5298-2862-4650-950d-abd7f8195942-cert\") pod \"openstack-operator-controller-manager-6f95c9794b-jtlvc\" (UID: \"fd4f5298-2862-4650-950d-abd7f8195942\") " pod="openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc" Oct 10 13:30:55 crc kubenswrapper[4745]: E1010 13:30:55.749767 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-khqb8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f4d5dfdc6-qzh4v_openstack-operators(e6828db3-0a2c-4b98-8cbf-1eba04f13bc5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 10 13:30:55 crc kubenswrapper[4745]: E1010 13:30:55.880427 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-lr987" podUID="ba76d6db-10cc-4d7d-b620-e37533730395" Oct 10 13:30:55 crc kubenswrapper[4745]: I1010 13:30:55.922122 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc" Oct 10 13:30:55 crc kubenswrapper[4745]: E1010 13:30:55.924340 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-rz48f" podUID="0ec7257b-b487-453c-88e8-d279184acdfc" Oct 10 13:30:56 crc kubenswrapper[4745]: E1010 13:30:56.003355 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-b7w5k" podUID="36db8827-b9af-4b9b-87aa-3cfa7d005ee0" Oct 10 13:30:56 crc kubenswrapper[4745]: E1010 13:30:56.062970 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-96mn7" podUID="6d02fe8c-f0b4-4ef1-8e57-569a836708cd" Oct 10 13:30:56 crc kubenswrapper[4745]: E1010 13:30:56.147699 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qzh4v" podUID="e6828db3-0a2c-4b98-8cbf-1eba04f13bc5" Oct 10 13:30:56 crc kubenswrapper[4745]: E1010 13:30:56.168985 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2dmpc" podUID="83b0fe55-090f-448e-95c2-1d3cbbc1b54c" Oct 10 13:30:56 crc kubenswrapper[4745]: E1010 13:30:56.181017 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-xxvtr" podUID="e1c2fd77-699a-4c79-9c54-9251c5ad7d17" Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.540021 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc"] Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.722906 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-b7w5k" event={"ID":"36db8827-b9af-4b9b-87aa-3cfa7d005ee0","Type":"ContainerStarted","Data":"965197e2d1f33367a2fae55c1ecc603cad81580d2b3c22252c87ce72118883e6"} Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.723221 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-b7w5k" event={"ID":"36db8827-b9af-4b9b-87aa-3cfa7d005ee0","Type":"ContainerStarted","Data":"d087601927a62942f18fae83379abba32b5cd19676c7ad2e256ed1bddc18ce91"} Oct 10 13:30:56 crc kubenswrapper[4745]: E1010 13:30:56.724705 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-b7w5k" podUID="36db8827-b9af-4b9b-87aa-3cfa7d005ee0" Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.726228 4745 generic.go:334] "Generic (PLEG): container finished" podID="15e38293-23c6-4705-9aaf-1de10eb3fe1e" containerID="3e1821ff9fdea1c4124ad2bbbff87def5b1aaee08a7c447fe7b6ec0de9b06555" exitCode=0 Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.726598 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpb4j" event={"ID":"15e38293-23c6-4705-9aaf-1de10eb3fe1e","Type":"ContainerDied","Data":"3e1821ff9fdea1c4124ad2bbbff87def5b1aaee08a7c447fe7b6ec0de9b06555"} Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.726630 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpb4j" event={"ID":"15e38293-23c6-4705-9aaf-1de10eb3fe1e","Type":"ContainerStarted","Data":"3e67dbbf4d4605ccd256431ab48030f3d2fa3b5333a8351511881f0dc2632adb"} Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.732743 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-h4jn2" event={"ID":"a5094dca-dde7-4144-bfbc-b4f90c170a21","Type":"ContainerStarted","Data":"fb7f5ab48460b5d3d9eb38315b5d6f298131069dfa174cc8a232e0e4f0d01970"} Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.736218 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-lr987" event={"ID":"ba76d6db-10cc-4d7d-b620-e37533730395","Type":"ContainerStarted","Data":"fc082a0d507d54d602a0f0a193a1051253fdc4e8fd758c6eb1b9f814a476016c"} Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.736246 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-lr987" event={"ID":"ba76d6db-10cc-4d7d-b620-e37533730395","Type":"ContainerStarted","Data":"5acf01f624ebf23fcab4992b34eb6ea58a6e2d218b19beeeccad320c3195986a"} Oct 10 13:30:56 crc kubenswrapper[4745]: E1010 13:30:56.737638 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a\\\"\"" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-lr987" podUID="ba76d6db-10cc-4d7d-b620-e37533730395" Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.737803 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc" event={"ID":"fd4f5298-2862-4650-950d-abd7f8195942","Type":"ContainerStarted","Data":"b33c443ec05e59d03aef331099d6fed87a1a5afd9a51c753e48ef9beb8c51aa9"} Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.740582 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2dmpc" event={"ID":"83b0fe55-090f-448e-95c2-1d3cbbc1b54c","Type":"ContainerStarted","Data":"5af39e135982abfff45051755ef95acdaa5616c0d472f5a8e66f3d079941fdd3"} Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.740610 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2dmpc" event={"ID":"83b0fe55-090f-448e-95c2-1d3cbbc1b54c","Type":"ContainerStarted","Data":"256cfb52d9261c0456ca89e2ff50a6f9f2ee90211cb08be1e3462fdfae7ee4a8"} Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.742222 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx" event={"ID":"443416f7-e4ce-41f2-b546-9344f0cb72a7","Type":"ContainerStarted","Data":"48ec6bc9f2f512fd544daaecfa0e509a9bc4fd29b8d858f1fd6cdac8617b121f"} Oct 10 13:30:56 crc kubenswrapper[4745]: E1010 13:30:56.743875 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2dmpc" podUID="83b0fe55-090f-448e-95c2-1d3cbbc1b54c" Oct 10 13:30:56 crc kubenswrapper[4745]: E1010 13:30:56.763932 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qzh4v" podUID="e6828db3-0a2c-4b98-8cbf-1eba04f13bc5" Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.765669 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dd7d5df-eee1-4f07-8dd8-24873ec41827" path="/var/lib/kubelet/pods/2dd7d5df-eee1-4f07-8dd8-24873ec41827/volumes" Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.766342 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qzh4v" event={"ID":"e6828db3-0a2c-4b98-8cbf-1eba04f13bc5","Type":"ContainerStarted","Data":"ff8d7ca39d44e2e24736fe5cffab19046d078132a8006bda7418fa3109bdd5dc"} Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.766375 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qzh4v" event={"ID":"e6828db3-0a2c-4b98-8cbf-1eba04f13bc5","Type":"ContainerStarted","Data":"9979c7db611db9eb8d330c469ee46d2fcfc551ab38dbe0e8ca4e5c5e5bf887ec"} Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.766386 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-96mn7" event={"ID":"6d02fe8c-f0b4-4ef1-8e57-569a836708cd","Type":"ContainerStarted","Data":"eaf3c5cf57bd2e9524a323689e0c5f4ac76a53ab8c37ca3c7789c4d13e44f093"} Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.766396 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-96mn7" event={"ID":"6d02fe8c-f0b4-4ef1-8e57-569a836708cd","Type":"ContainerStarted","Data":"8b0d71ab75944725b535188ef6106d1e69c4877e2236b34c54f9dd55eb0747b9"} Oct 10 13:30:56 crc kubenswrapper[4745]: E1010 13:30:56.772418 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-96mn7" podUID="6d02fe8c-f0b4-4ef1-8e57-569a836708cd" Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.785122 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw" event={"ID":"0174cdbc-06e0-423f-af63-5467737e6ec5","Type":"ContainerStarted","Data":"f8b81bc773e4c9f018eed2a1dc8927d13bf2d5c3ac789035e9f5fbcebe7a74a0"} Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.796715 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c9gmp" event={"ID":"ebc60df2-d5b0-418f-a8cd-e8440706d3be","Type":"ContainerStarted","Data":"70865a34ad34073933059dd3d3cc8f45494cefd3726d6b208b0b4c0436318a4b"} Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.799159 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-xxvtr" event={"ID":"e1c2fd77-699a-4c79-9c54-9251c5ad7d17","Type":"ContainerStarted","Data":"bf778023e93dcc7e1f873dbaeddc7ac486d44332bf707640d60a7f5e1ddabaad"} Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.799213 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-xxvtr" event={"ID":"e1c2fd77-699a-4c79-9c54-9251c5ad7d17","Type":"ContainerStarted","Data":"9dc57b17b993edf9c4039bd15aa3a969549ea3b1bb170c3da1b5a0578e4e0178"} Oct 10 13:30:56 crc kubenswrapper[4745]: E1010 13:30:56.805644 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:09deecf840d38ff6af3c924729cf0a9444bc985848bfbe7c918019b88a6bc4d7\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-xxvtr" podUID="e1c2fd77-699a-4c79-9c54-9251c5ad7d17" Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.809635 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-rz48f" event={"ID":"0ec7257b-b487-453c-88e8-d279184acdfc","Type":"ContainerStarted","Data":"18bcb25ef1b4adb94417cc42ce1e9ab01afc46725e0d824cf56348bb10563332"} Oct 10 13:30:56 crc kubenswrapper[4745]: I1010 13:30:56.809683 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-rz48f" event={"ID":"0ec7257b-b487-453c-88e8-d279184acdfc","Type":"ContainerStarted","Data":"599c7c2c033a7a5422b4a3831d17452eb9b8ee448e5714fa9e700a5c059a8dc5"} Oct 10 13:30:56 crc kubenswrapper[4745]: E1010 13:30:56.811490 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-rz48f" podUID="0ec7257b-b487-453c-88e8-d279184acdfc" Oct 10 13:30:57 crc kubenswrapper[4745]: I1010 13:30:57.832088 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc" event={"ID":"fd4f5298-2862-4650-950d-abd7f8195942","Type":"ContainerStarted","Data":"2cc151fdd0acb37447955cc4d94f7ea00dc708957ed7d0615b6f215ba0de889c"} Oct 10 13:30:57 crc kubenswrapper[4745]: I1010 13:30:57.832405 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc" event={"ID":"fd4f5298-2862-4650-950d-abd7f8195942","Type":"ContainerStarted","Data":"c3ac5efcf57fd9250b4a65cc1ad7dc3d00c689271ea5d406c56460444d886058"} Oct 10 13:30:57 crc kubenswrapper[4745]: E1010 13:30:57.835137 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-96mn7" podUID="6d02fe8c-f0b4-4ef1-8e57-569a836708cd" Oct 10 13:30:57 crc kubenswrapper[4745]: E1010 13:30:57.835421 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qzh4v" podUID="e6828db3-0a2c-4b98-8cbf-1eba04f13bc5" Oct 10 13:30:57 crc kubenswrapper[4745]: I1010 13:30:57.835558 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc" Oct 10 13:30:57 crc kubenswrapper[4745]: E1010 13:30:57.852041 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a\\\"\"" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-lr987" podUID="ba76d6db-10cc-4d7d-b620-e37533730395" Oct 10 13:30:57 crc kubenswrapper[4745]: E1010 13:30:57.852314 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:09deecf840d38ff6af3c924729cf0a9444bc985848bfbe7c918019b88a6bc4d7\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-xxvtr" podUID="e1c2fd77-699a-4c79-9c54-9251c5ad7d17" Oct 10 13:30:57 crc kubenswrapper[4745]: E1010 13:30:57.852365 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-rz48f" podUID="0ec7257b-b487-453c-88e8-d279184acdfc" Oct 10 13:30:57 crc kubenswrapper[4745]: E1010 13:30:57.852550 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-b7w5k" podUID="36db8827-b9af-4b9b-87aa-3cfa7d005ee0" Oct 10 13:30:57 crc kubenswrapper[4745]: E1010 13:30:57.854740 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2dmpc" podUID="83b0fe55-090f-448e-95c2-1d3cbbc1b54c" Oct 10 13:30:57 crc kubenswrapper[4745]: I1010 13:30:57.941955 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9jkxn" Oct 10 13:30:57 crc kubenswrapper[4745]: I1010 13:30:57.943094 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9jkxn" Oct 10 13:30:57 crc kubenswrapper[4745]: I1010 13:30:57.986640 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc" podStartSLOduration=4.986620962 podStartE2EDuration="4.986620962s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:30:57.981251699 +0000 UTC m=+771.878908462" watchObservedRunningTime="2025-10-10 13:30:57.986620962 +0000 UTC m=+771.884277725" Oct 10 13:30:58 crc kubenswrapper[4745]: I1010 13:30:58.073974 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9jkxn" Oct 10 13:30:58 crc kubenswrapper[4745]: I1010 13:30:58.875325 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9jkxn" Oct 10 13:31:00 crc kubenswrapper[4745]: I1010 13:31:00.400980 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9jkxn"] Oct 10 13:31:01 crc kubenswrapper[4745]: I1010 13:31:01.857470 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9jkxn" podUID="8e605d21-bc5f-44ee-a3f7-7558d307162b" containerName="registry-server" containerID="cri-o://136dcf058e7d5d64edee0c76f38356fb2bbf311dcc3d135959ac715111cfa951" gracePeriod=2 Oct 10 13:31:02 crc kubenswrapper[4745]: I1010 13:31:02.869870 4745 generic.go:334] "Generic (PLEG): container finished" podID="8e605d21-bc5f-44ee-a3f7-7558d307162b" containerID="136dcf058e7d5d64edee0c76f38356fb2bbf311dcc3d135959ac715111cfa951" exitCode=0 Oct 10 13:31:02 crc kubenswrapper[4745]: I1010 13:31:02.869944 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jkxn" event={"ID":"8e605d21-bc5f-44ee-a3f7-7558d307162b","Type":"ContainerDied","Data":"136dcf058e7d5d64edee0c76f38356fb2bbf311dcc3d135959ac715111cfa951"} Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.563242 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9jkxn" Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.688634 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e605d21-bc5f-44ee-a3f7-7558d307162b-utilities\") pod \"8e605d21-bc5f-44ee-a3f7-7558d307162b\" (UID: \"8e605d21-bc5f-44ee-a3f7-7558d307162b\") " Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.688719 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxn5z\" (UniqueName: \"kubernetes.io/projected/8e605d21-bc5f-44ee-a3f7-7558d307162b-kube-api-access-rxn5z\") pod \"8e605d21-bc5f-44ee-a3f7-7558d307162b\" (UID: \"8e605d21-bc5f-44ee-a3f7-7558d307162b\") " Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.688795 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e605d21-bc5f-44ee-a3f7-7558d307162b-catalog-content\") pod \"8e605d21-bc5f-44ee-a3f7-7558d307162b\" (UID: \"8e605d21-bc5f-44ee-a3f7-7558d307162b\") " Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.689689 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e605d21-bc5f-44ee-a3f7-7558d307162b-utilities" (OuterVolumeSpecName: "utilities") pod "8e605d21-bc5f-44ee-a3f7-7558d307162b" (UID: "8e605d21-bc5f-44ee-a3f7-7558d307162b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.694982 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e605d21-bc5f-44ee-a3f7-7558d307162b-kube-api-access-rxn5z" (OuterVolumeSpecName: "kube-api-access-rxn5z") pod "8e605d21-bc5f-44ee-a3f7-7558d307162b" (UID: "8e605d21-bc5f-44ee-a3f7-7558d307162b"). InnerVolumeSpecName "kube-api-access-rxn5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.707679 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e605d21-bc5f-44ee-a3f7-7558d307162b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e605d21-bc5f-44ee-a3f7-7558d307162b" (UID: "8e605d21-bc5f-44ee-a3f7-7558d307162b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.789851 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxn5z\" (UniqueName: \"kubernetes.io/projected/8e605d21-bc5f-44ee-a3f7-7558d307162b-kube-api-access-rxn5z\") on node \"crc\" DevicePath \"\"" Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.790251 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e605d21-bc5f-44ee-a3f7-7558d307162b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.790265 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e605d21-bc5f-44ee-a3f7-7558d307162b-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.894667 4745 generic.go:334] "Generic (PLEG): container finished" podID="15e38293-23c6-4705-9aaf-1de10eb3fe1e" containerID="9a212e58e6524c1210f213d705609b9e98b98e82dbadc40442173cede1aedda3" exitCode=0 Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.894783 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpb4j" event={"ID":"15e38293-23c6-4705-9aaf-1de10eb3fe1e","Type":"ContainerDied","Data":"9a212e58e6524c1210f213d705609b9e98b98e82dbadc40442173cede1aedda3"} Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.898749 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jkxn" event={"ID":"8e605d21-bc5f-44ee-a3f7-7558d307162b","Type":"ContainerDied","Data":"0f659751771f20c1317b8363985766299f9773610eac060c31094c9f8aa40ca7"} Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.898805 4745 scope.go:117] "RemoveContainer" containerID="136dcf058e7d5d64edee0c76f38356fb2bbf311dcc3d135959ac715111cfa951" Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.898846 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9jkxn" Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.932590 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6f95c9794b-jtlvc" Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.967954 4745 scope.go:117] "RemoveContainer" containerID="de2ca6f748b81cd399df67160f038b7a52277e5e3b0efb0b149cd3ee230681bc" Oct 10 13:31:05 crc kubenswrapper[4745]: I1010 13:31:05.995573 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9jkxn"] Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.010642 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9jkxn"] Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.023815 4745 scope.go:117] "RemoveContainer" containerID="856d6a9fa1720c9f341eb44b43ad53ec1793fbbc7d6e925b18a0fd00788c7aa8" Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.791215 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e605d21-bc5f-44ee-a3f7-7558d307162b" path="/var/lib/kubelet/pods/8e605d21-bc5f-44ee-a3f7-7558d307162b/volumes" Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.923132 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-fpcxk" event={"ID":"dff847b8-915b-4b64-85c3-d7c20d1282a0","Type":"ContainerStarted","Data":"7abf4c09b8a508398834194396d7f58f08a6d314544d2dd9cd12a9a97c59d592"} Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.923174 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-fpcxk" event={"ID":"dff847b8-915b-4b64-85c3-d7c20d1282a0","Type":"ContainerStarted","Data":"eb5a033866c237a25ac9a60c00f7d527d3fa1e9be9695671fec3029917f0ec4d"} Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.923644 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-fpcxk" Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.925927 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-t9wqw" event={"ID":"fc401fc1-e03b-4b32-8782-15e9c2a70e8a","Type":"ContainerStarted","Data":"87296becbb9521493212ab303f1d2ea6aa2ded32656bf21cc02ea880284108a9"} Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.925950 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-t9wqw" event={"ID":"fc401fc1-e03b-4b32-8782-15e9c2a70e8a","Type":"ContainerStarted","Data":"aa25b8b57994c2f39028bf1d2b768edd58840da9a25dea08e2659c736a007f15"} Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.926295 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-59578bc799-t9wqw" Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.933072 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx" event={"ID":"443416f7-e4ce-41f2-b546-9344f0cb72a7","Type":"ContainerStarted","Data":"649191ca9c9c34a753bd0d381c46e67cc6b4e83d3b8bfcfc93006d8d6ab6008c"} Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.946961 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-89lzv" event={"ID":"93defbbc-c732-4743-bcff-a1d7f06aefa2","Type":"ContainerStarted","Data":"f245a31a02bb8745cd3f0a496e5ab225762751c0e311a19bfde8dea913a16ed0"} Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.947024 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-89lzv" event={"ID":"93defbbc-c732-4743-bcff-a1d7f06aefa2","Type":"ContainerStarted","Data":"dc033ce477720d0c2f92c4144e8e5d8d38b0cfa8bbf8f99f083da90ea4a7afae"} Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.947466 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-89lzv" Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.948523 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-fpcxk" podStartSLOduration=3.329433049 podStartE2EDuration="13.948507317s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:55.004870855 +0000 UTC m=+768.902527618" lastFinishedPulling="2025-10-10 13:31:05.623945103 +0000 UTC m=+779.521601886" observedRunningTime="2025-10-10 13:31:06.946353477 +0000 UTC m=+780.844010240" watchObservedRunningTime="2025-10-10 13:31:06.948507317 +0000 UTC m=+780.846164080" Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.960112 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-c7986" event={"ID":"5ca519af-249e-48f0-8669-77bd548be8f3","Type":"ContainerStarted","Data":"4c955c409a1de36bb0871f1f4010737de567e2ae38db8525fac0717cfdd82197"} Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.972249 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-59578bc799-t9wqw" podStartSLOduration=3.3498073870000002 podStartE2EDuration="13.972228681s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:55.002426979 +0000 UTC m=+768.900083742" lastFinishedPulling="2025-10-10 13:31:05.624848253 +0000 UTC m=+779.522505036" observedRunningTime="2025-10-10 13:31:06.963286416 +0000 UTC m=+780.860943179" watchObservedRunningTime="2025-10-10 13:31:06.972228681 +0000 UTC m=+780.869885444" Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.985275 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-89lzv" podStartSLOduration=3.495408866 podStartE2EDuration="13.985248599s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:55.125587294 +0000 UTC m=+769.023244047" lastFinishedPulling="2025-10-10 13:31:05.615427017 +0000 UTC m=+779.513083780" observedRunningTime="2025-10-10 13:31:06.985162197 +0000 UTC m=+780.882818960" watchObservedRunningTime="2025-10-10 13:31:06.985248599 +0000 UTC m=+780.882905362" Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.988270 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw" event={"ID":"0174cdbc-06e0-423f-af63-5467737e6ec5","Type":"ContainerStarted","Data":"072931cc6a49f04af734b9a27b10c1b14d3a06e98661ce780166ea04fa738151"} Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.988323 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw" event={"ID":"0174cdbc-06e0-423f-af63-5467737e6ec5","Type":"ContainerStarted","Data":"038e65a032b179175703872fe7f4f46f11874a2594d8f3403eeebd5fcc9b0fb1"} Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.989225 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw" Oct 10 13:31:06 crc kubenswrapper[4745]: I1010 13:31:06.998782 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-d5qnn" event={"ID":"9fdf7ed4-1e18-43c0-b305-2b9b226f4634","Type":"ContainerStarted","Data":"338e745dcd895979f177b1c789bc895585ca9aa9acda40108aedb03f599413de"} Oct 10 13:31:07 crc kubenswrapper[4745]: I1010 13:31:07.004384 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-h2bt5" event={"ID":"10978ce4-0034-4ad4-9a47-2f1931c09655","Type":"ContainerStarted","Data":"83fac6bb6cf0a8603faf65a833822f2953e0ec8c1302bf0bb71ec89d383b70b4"} Oct 10 13:31:07 crc kubenswrapper[4745]: I1010 13:31:07.013640 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tz665" event={"ID":"446e0b63-81e5-47fb-b110-cfa5393e5311","Type":"ContainerStarted","Data":"fc3acd852e9d3cfacd761089a47d211314eb576be7ef580e07cce9f78bb6e73a"} Oct 10 13:31:07 crc kubenswrapper[4745]: I1010 13:31:07.018136 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-2w6nw" event={"ID":"7f07678c-0cc2-47a5-b244-72ae3aa55baa","Type":"ContainerStarted","Data":"fd2fbcf35fe08e0e393dfec6ed9f8f2cea4060a907da0659e7e4d931f8b48f40"} Oct 10 13:31:07 crc kubenswrapper[4745]: I1010 13:31:07.027019 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-vc9rk" event={"ID":"420fdd11-0131-479d-9f7a-47f091cacd83","Type":"ContainerStarted","Data":"be378d03a596f6c4197edbb7dfceba4c1b313c5e7e266f76979b8c04ee26a8d2"} Oct 10 13:31:07 crc kubenswrapper[4745]: I1010 13:31:07.037553 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c9gmp" event={"ID":"ebc60df2-d5b0-418f-a8cd-e8440706d3be","Type":"ContainerStarted","Data":"0e0d09d5ebc8623a1a9506972079e066fd876182f038a0a02f7bb0be3b88aff8"} Oct 10 13:31:07 crc kubenswrapper[4745]: I1010 13:31:07.039627 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-h4jn2" event={"ID":"a5094dca-dde7-4144-bfbc-b4f90c170a21","Type":"ContainerStarted","Data":"671369091d7faede131f8111d5d53ba95162075ab3bd9cffd43a557cb52e1fcf"} Oct 10 13:31:07 crc kubenswrapper[4745]: I1010 13:31:07.040660 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-htmcv" event={"ID":"730b5555-5f6d-4e16-91e0-faacad56d777","Type":"ContainerStarted","Data":"52af0c2ad2effd3d3f9d02179740762918a4b3a423b6491749f088b9b13caaa6"} Oct 10 13:31:07 crc kubenswrapper[4745]: I1010 13:31:07.040694 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-htmcv" event={"ID":"730b5555-5f6d-4e16-91e0-faacad56d777","Type":"ContainerStarted","Data":"a8e7a737a4a16248ccfcff3d3cb42cf4cf138bd4761c44d53c13457ff84aa040"} Oct 10 13:31:07 crc kubenswrapper[4745]: I1010 13:31:07.041191 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-htmcv" Oct 10 13:31:07 crc kubenswrapper[4745]: I1010 13:31:07.069336 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw" podStartSLOduration=4.237331095 podStartE2EDuration="14.069297797s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:55.789226467 +0000 UTC m=+769.686883230" lastFinishedPulling="2025-10-10 13:31:05.621193169 +0000 UTC m=+779.518849932" observedRunningTime="2025-10-10 13:31:07.042238927 +0000 UTC m=+780.939895690" watchObservedRunningTime="2025-10-10 13:31:07.069297797 +0000 UTC m=+780.966954730" Oct 10 13:31:07 crc kubenswrapper[4745]: I1010 13:31:07.075245 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-qmbvh" event={"ID":"1d2b124e-5495-42bd-b675-4eed70ccc446","Type":"ContainerStarted","Data":"b89ba008a4ae40a62ad4a0480c058e6bc8401a499accde1e068e08ac8796537c"} Oct 10 13:31:07 crc kubenswrapper[4745]: I1010 13:31:07.075307 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-qmbvh" event={"ID":"1d2b124e-5495-42bd-b675-4eed70ccc446","Type":"ContainerStarted","Data":"bb607f4722a2920b34b9fcc78995147922eca302ac6a2889ae3b5ebf1235246b"} Oct 10 13:31:07 crc kubenswrapper[4745]: I1010 13:31:07.075643 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-qmbvh" Oct 10 13:31:07 crc kubenswrapper[4745]: I1010 13:31:07.139764 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c9gmp" podStartSLOduration=4.08190188 podStartE2EDuration="14.139725653s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:55.534061154 +0000 UTC m=+769.431717917" lastFinishedPulling="2025-10-10 13:31:05.591884907 +0000 UTC m=+779.489541690" observedRunningTime="2025-10-10 13:31:07.094566367 +0000 UTC m=+780.992223130" watchObservedRunningTime="2025-10-10 13:31:07.139725653 +0000 UTC m=+781.037382416" Oct 10 13:31:07 crc kubenswrapper[4745]: I1010 13:31:07.143290 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-htmcv" podStartSLOduration=2.878031925 podStartE2EDuration="14.143281775s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:54.29885617 +0000 UTC m=+768.196512923" lastFinishedPulling="2025-10-10 13:31:05.56410601 +0000 UTC m=+779.461762773" observedRunningTime="2025-10-10 13:31:07.141494144 +0000 UTC m=+781.039150907" watchObservedRunningTime="2025-10-10 13:31:07.143281775 +0000 UTC m=+781.040938538" Oct 10 13:31:07 crc kubenswrapper[4745]: I1010 13:31:07.167469 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-qmbvh" podStartSLOduration=3.784800035 podStartE2EDuration="14.167455659s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:55.16289308 +0000 UTC m=+769.060549833" lastFinishedPulling="2025-10-10 13:31:05.545548684 +0000 UTC m=+779.443205457" observedRunningTime="2025-10-10 13:31:07.162835933 +0000 UTC m=+781.060492706" watchObservedRunningTime="2025-10-10 13:31:07.167455659 +0000 UTC m=+781.065112422" Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.102830 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-h4jn2" event={"ID":"a5094dca-dde7-4144-bfbc-b4f90c170a21","Type":"ContainerStarted","Data":"c672560404bd3a048c4f9c2d5894e6eecf9795e5ff62ed6e0f284f3ec1a27539"} Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.104399 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-h4jn2" Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.106218 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-2w6nw" event={"ID":"7f07678c-0cc2-47a5-b244-72ae3aa55baa","Type":"ContainerStarted","Data":"66551d99d3957b31d18e0343ddd6be307213ab45845c77121fbd4bb8694afe45"} Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.106583 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-2w6nw" Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.108217 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tz665" event={"ID":"446e0b63-81e5-47fb-b110-cfa5393e5311","Type":"ContainerStarted","Data":"af2d79a038a3987ab1f5dbc3a8407231d54a3f6781df95fc21a0b6a298a58158"} Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.108365 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tz665" Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.110167 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-vc9rk" event={"ID":"420fdd11-0131-479d-9f7a-47f091cacd83","Type":"ContainerStarted","Data":"ccec91d8059276f45e29a3136e6ddd4db5ee50d7b224b0e9935ad25885a7e799"} Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.110551 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-vc9rk" Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.115820 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx" event={"ID":"443416f7-e4ce-41f2-b546-9344f0cb72a7","Type":"ContainerStarted","Data":"1386998870bfca9dab48e0fe59f65ebd66916bf90a355d6208e6057ac6e49aa4"} Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.116237 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx" Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.122077 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-c7986" event={"ID":"5ca519af-249e-48f0-8669-77bd548be8f3","Type":"ContainerStarted","Data":"0c2a8f74226a77b4d90510b52e62424f5facd1e65f0572548fe229829c9c2612"} Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.123155 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-c7986" Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.126126 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpb4j" event={"ID":"15e38293-23c6-4705-9aaf-1de10eb3fe1e","Type":"ContainerStarted","Data":"957cd8f8a258815fcc84f7340096c353808994c13d0dcdbab2d7a4ca1da38d64"} Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.126979 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-h4jn2" podStartSLOduration=5.058145704 podStartE2EDuration="15.126962869s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:55.559387125 +0000 UTC m=+769.457043888" lastFinishedPulling="2025-10-10 13:31:05.62820429 +0000 UTC m=+779.525861053" observedRunningTime="2025-10-10 13:31:08.121450793 +0000 UTC m=+782.019107556" watchObservedRunningTime="2025-10-10 13:31:08.126962869 +0000 UTC m=+782.024619632" Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.128114 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-d5qnn" event={"ID":"9fdf7ed4-1e18-43c0-b305-2b9b226f4634","Type":"ContainerStarted","Data":"6ae668475d9e89cd8d35d7d2f630ebe69564ffba21960b39c142d99a961a3bf1"} Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.128316 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-d5qnn" Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.130239 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-h2bt5" event={"ID":"10978ce4-0034-4ad4-9a47-2f1931c09655","Type":"ContainerStarted","Data":"bb722d1070e8d634e56e199c54f3562163c7c2dbd90b39d528715ebb19072d79"} Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.130681 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-h2bt5" Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.155810 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx" podStartSLOduration=5.061015989 podStartE2EDuration="15.15578242s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:55.559914237 +0000 UTC m=+769.457571000" lastFinishedPulling="2025-10-10 13:31:05.654680668 +0000 UTC m=+779.552337431" observedRunningTime="2025-10-10 13:31:08.145668418 +0000 UTC m=+782.043325191" watchObservedRunningTime="2025-10-10 13:31:08.15578242 +0000 UTC m=+782.053439223" Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.175063 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tz665" podStartSLOduration=4.528614887 podStartE2EDuration="15.175038272s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:54.976880013 +0000 UTC m=+768.874536776" lastFinishedPulling="2025-10-10 13:31:05.623303398 +0000 UTC m=+779.520960161" observedRunningTime="2025-10-10 13:31:08.168301597 +0000 UTC m=+782.065958390" watchObservedRunningTime="2025-10-10 13:31:08.175038272 +0000 UTC m=+782.072695035" Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.187424 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-vc9rk" podStartSLOduration=4.815495668 podStartE2EDuration="15.187404406s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:55.182785556 +0000 UTC m=+769.080442319" lastFinishedPulling="2025-10-10 13:31:05.554694284 +0000 UTC m=+779.452351057" observedRunningTime="2025-10-10 13:31:08.184890358 +0000 UTC m=+782.082547121" watchObservedRunningTime="2025-10-10 13:31:08.187404406 +0000 UTC m=+782.085061169" Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.204960 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-2w6nw" podStartSLOduration=4.5313627709999995 podStartE2EDuration="15.204933668s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:54.954023899 +0000 UTC m=+768.851680662" lastFinishedPulling="2025-10-10 13:31:05.627594796 +0000 UTC m=+779.525251559" observedRunningTime="2025-10-10 13:31:08.200825414 +0000 UTC m=+782.098482177" watchObservedRunningTime="2025-10-10 13:31:08.204933668 +0000 UTC m=+782.102590451" Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.219198 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-h2bt5" podStartSLOduration=4.586995537 podStartE2EDuration="15.219181345s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:54.991968079 +0000 UTC m=+768.889624842" lastFinishedPulling="2025-10-10 13:31:05.624153887 +0000 UTC m=+779.521810650" observedRunningTime="2025-10-10 13:31:08.217395654 +0000 UTC m=+782.115052427" watchObservedRunningTime="2025-10-10 13:31:08.219181345 +0000 UTC m=+782.116838108" Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.240902 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-c7986" podStartSLOduration=4.678046355 podStartE2EDuration="15.240884842s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:54.991919208 +0000 UTC m=+768.889575971" lastFinishedPulling="2025-10-10 13:31:05.554757695 +0000 UTC m=+779.452414458" observedRunningTime="2025-10-10 13:31:08.232382417 +0000 UTC m=+782.130039180" watchObservedRunningTime="2025-10-10 13:31:08.240884842 +0000 UTC m=+782.138541605" Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.265215 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tpb4j" podStartSLOduration=4.473568863 podStartE2EDuration="14.26519888s" podCreationTimestamp="2025-10-10 13:30:54 +0000 UTC" firstStartedPulling="2025-10-10 13:30:56.729318752 +0000 UTC m=+770.626975515" lastFinishedPulling="2025-10-10 13:31:06.520948779 +0000 UTC m=+780.418605532" observedRunningTime="2025-10-10 13:31:08.260511513 +0000 UTC m=+782.158168276" watchObservedRunningTime="2025-10-10 13:31:08.26519888 +0000 UTC m=+782.162855643" Oct 10 13:31:08 crc kubenswrapper[4745]: I1010 13:31:08.283566 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-d5qnn" podStartSLOduration=4.635290834 podStartE2EDuration="15.283546621s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:54.980026585 +0000 UTC m=+768.877683348" lastFinishedPulling="2025-10-10 13:31:05.628282362 +0000 UTC m=+779.525939135" observedRunningTime="2025-10-10 13:31:08.281129976 +0000 UTC m=+782.178786759" watchObservedRunningTime="2025-10-10 13:31:08.283546621 +0000 UTC m=+782.181203374" Oct 10 13:31:11 crc kubenswrapper[4745]: I1010 13:31:11.184699 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-lr987" event={"ID":"ba76d6db-10cc-4d7d-b620-e37533730395","Type":"ContainerStarted","Data":"35694159373c6be22024aacf206a44c6dd2f36d4dcc0203ea527d308f8c2a3f7"} Oct 10 13:31:11 crc kubenswrapper[4745]: I1010 13:31:11.185929 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-lr987" Oct 10 13:31:11 crc kubenswrapper[4745]: I1010 13:31:11.207100 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-lr987" podStartSLOduration=3.289970294 podStartE2EDuration="18.207079963s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:55.560196814 +0000 UTC m=+769.457853577" lastFinishedPulling="2025-10-10 13:31:10.477306443 +0000 UTC m=+784.374963246" observedRunningTime="2025-10-10 13:31:11.20392204 +0000 UTC m=+785.101578803" watchObservedRunningTime="2025-10-10 13:31:11.207079963 +0000 UTC m=+785.104736716" Oct 10 13:31:13 crc kubenswrapper[4745]: I1010 13:31:13.533985 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-htmcv" Oct 10 13:31:13 crc kubenswrapper[4745]: I1010 13:31:13.571466 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-c7986" Oct 10 13:31:13 crc kubenswrapper[4745]: I1010 13:31:13.592762 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-fpcxk" Oct 10 13:31:13 crc kubenswrapper[4745]: I1010 13:31:13.635532 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tz665" Oct 10 13:31:13 crc kubenswrapper[4745]: I1010 13:31:13.652456 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-2w6nw" Oct 10 13:31:13 crc kubenswrapper[4745]: I1010 13:31:13.689203 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-59578bc799-t9wqw" Oct 10 13:31:13 crc kubenswrapper[4745]: I1010 13:31:13.703217 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-qmbvh" Oct 10 13:31:13 crc kubenswrapper[4745]: I1010 13:31:13.820644 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-d5qnn" Oct 10 13:31:13 crc kubenswrapper[4745]: I1010 13:31:13.850280 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-h2bt5" Oct 10 13:31:14 crc kubenswrapper[4745]: I1010 13:31:14.002809 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-vc9rk" Oct 10 13:31:14 crc kubenswrapper[4745]: I1010 13:31:14.047582 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-89lzv" Oct 10 13:31:14 crc kubenswrapper[4745]: I1010 13:31:14.309678 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-h4jn2" Oct 10 13:31:14 crc kubenswrapper[4745]: I1010 13:31:14.376079 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-dt9kx" Oct 10 13:31:14 crc kubenswrapper[4745]: I1010 13:31:14.687938 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw" Oct 10 13:31:15 crc kubenswrapper[4745]: I1010 13:31:15.136887 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tpb4j" Oct 10 13:31:15 crc kubenswrapper[4745]: I1010 13:31:15.136948 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tpb4j" Oct 10 13:31:15 crc kubenswrapper[4745]: I1010 13:31:15.204004 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tpb4j" Oct 10 13:31:15 crc kubenswrapper[4745]: I1010 13:31:15.309965 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tpb4j" Oct 10 13:31:15 crc kubenswrapper[4745]: I1010 13:31:15.453919 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tpb4j"] Oct 10 13:31:17 crc kubenswrapper[4745]: I1010 13:31:17.241972 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tpb4j" podUID="15e38293-23c6-4705-9aaf-1de10eb3fe1e" containerName="registry-server" containerID="cri-o://957cd8f8a258815fcc84f7340096c353808994c13d0dcdbab2d7a4ca1da38d64" gracePeriod=2 Oct 10 13:31:17 crc kubenswrapper[4745]: E1010 13:31:17.920984 4745 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15e38293_23c6_4705_9aaf_1de10eb3fe1e.slice/crio-conmon-957cd8f8a258815fcc84f7340096c353808994c13d0dcdbab2d7a4ca1da38d64.scope\": RecentStats: unable to find data in memory cache]" Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.254867 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-b7w5k" event={"ID":"36db8827-b9af-4b9b-87aa-3cfa7d005ee0","Type":"ContainerStarted","Data":"88a2cba5ee7769072f41606f1b91de55ec233ec1231fd38e87f242ed75a747b9"} Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.255417 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-664664cb68-b7w5k" Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.256706 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-96mn7" event={"ID":"6d02fe8c-f0b4-4ef1-8e57-569a836708cd","Type":"ContainerStarted","Data":"34614d8c836f1a34c549c450db091172ea77c6c7532acf4a955f20c57f439563"} Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.257013 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-646675d848-96mn7" Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.261169 4745 generic.go:334] "Generic (PLEG): container finished" podID="15e38293-23c6-4705-9aaf-1de10eb3fe1e" containerID="957cd8f8a258815fcc84f7340096c353808994c13d0dcdbab2d7a4ca1da38d64" exitCode=0 Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.261210 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpb4j" event={"ID":"15e38293-23c6-4705-9aaf-1de10eb3fe1e","Type":"ContainerDied","Data":"957cd8f8a258815fcc84f7340096c353808994c13d0dcdbab2d7a4ca1da38d64"} Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.277624 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-664664cb68-b7w5k" podStartSLOduration=3.061975135 podStartE2EDuration="25.277584453s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:55.585458713 +0000 UTC m=+769.483115476" lastFinishedPulling="2025-10-10 13:31:17.801067991 +0000 UTC m=+791.698724794" observedRunningTime="2025-10-10 13:31:18.277216474 +0000 UTC m=+792.174873267" watchObservedRunningTime="2025-10-10 13:31:18.277584453 +0000 UTC m=+792.175241236" Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.290648 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-646675d848-96mn7" podStartSLOduration=3.122492258 podStartE2EDuration="25.290636297s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:55.658074809 +0000 UTC m=+769.555731572" lastFinishedPulling="2025-10-10 13:31:17.826218808 +0000 UTC m=+791.723875611" observedRunningTime="2025-10-10 13:31:18.288592575 +0000 UTC m=+792.186249328" watchObservedRunningTime="2025-10-10 13:31:18.290636297 +0000 UTC m=+792.188293060" Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.746670 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tpb4j" Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.827474 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp545\" (UniqueName: \"kubernetes.io/projected/15e38293-23c6-4705-9aaf-1de10eb3fe1e-kube-api-access-cp545\") pod \"15e38293-23c6-4705-9aaf-1de10eb3fe1e\" (UID: \"15e38293-23c6-4705-9aaf-1de10eb3fe1e\") " Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.827643 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15e38293-23c6-4705-9aaf-1de10eb3fe1e-utilities\") pod \"15e38293-23c6-4705-9aaf-1de10eb3fe1e\" (UID: \"15e38293-23c6-4705-9aaf-1de10eb3fe1e\") " Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.827699 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15e38293-23c6-4705-9aaf-1de10eb3fe1e-catalog-content\") pod \"15e38293-23c6-4705-9aaf-1de10eb3fe1e\" (UID: \"15e38293-23c6-4705-9aaf-1de10eb3fe1e\") " Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.829061 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15e38293-23c6-4705-9aaf-1de10eb3fe1e-utilities" (OuterVolumeSpecName: "utilities") pod "15e38293-23c6-4705-9aaf-1de10eb3fe1e" (UID: "15e38293-23c6-4705-9aaf-1de10eb3fe1e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.833895 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15e38293-23c6-4705-9aaf-1de10eb3fe1e-kube-api-access-cp545" (OuterVolumeSpecName: "kube-api-access-cp545") pod "15e38293-23c6-4705-9aaf-1de10eb3fe1e" (UID: "15e38293-23c6-4705-9aaf-1de10eb3fe1e"). InnerVolumeSpecName "kube-api-access-cp545". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.876974 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15e38293-23c6-4705-9aaf-1de10eb3fe1e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15e38293-23c6-4705-9aaf-1de10eb3fe1e" (UID: "15e38293-23c6-4705-9aaf-1de10eb3fe1e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.929362 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp545\" (UniqueName: \"kubernetes.io/projected/15e38293-23c6-4705-9aaf-1de10eb3fe1e-kube-api-access-cp545\") on node \"crc\" DevicePath \"\"" Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.929391 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15e38293-23c6-4705-9aaf-1de10eb3fe1e-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:31:18 crc kubenswrapper[4745]: I1010 13:31:18.929405 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15e38293-23c6-4705-9aaf-1de10eb3fe1e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:31:19 crc kubenswrapper[4745]: I1010 13:31:19.273772 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpb4j" event={"ID":"15e38293-23c6-4705-9aaf-1de10eb3fe1e","Type":"ContainerDied","Data":"3e67dbbf4d4605ccd256431ab48030f3d2fa3b5333a8351511881f0dc2632adb"} Oct 10 13:31:19 crc kubenswrapper[4745]: I1010 13:31:19.273840 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tpb4j" Oct 10 13:31:19 crc kubenswrapper[4745]: I1010 13:31:19.273873 4745 scope.go:117] "RemoveContainer" containerID="957cd8f8a258815fcc84f7340096c353808994c13d0dcdbab2d7a4ca1da38d64" Oct 10 13:31:19 crc kubenswrapper[4745]: I1010 13:31:19.323611 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tpb4j"] Oct 10 13:31:19 crc kubenswrapper[4745]: I1010 13:31:19.331415 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tpb4j"] Oct 10 13:31:19 crc kubenswrapper[4745]: I1010 13:31:19.928970 4745 scope.go:117] "RemoveContainer" containerID="9a212e58e6524c1210f213d705609b9e98b98e82dbadc40442173cede1aedda3" Oct 10 13:31:19 crc kubenswrapper[4745]: I1010 13:31:19.954927 4745 scope.go:117] "RemoveContainer" containerID="3e1821ff9fdea1c4124ad2bbbff87def5b1aaee08a7c447fe7b6ec0de9b06555" Oct 10 13:31:20 crc kubenswrapper[4745]: I1010 13:31:20.282357 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2dmpc" event={"ID":"83b0fe55-090f-448e-95c2-1d3cbbc1b54c","Type":"ContainerStarted","Data":"0f3ddba1aa1c51bab11aa45f3d5f25c91e11d7e79aad18616a677fe0a61cb6ef"} Oct 10 13:31:20 crc kubenswrapper[4745]: I1010 13:31:20.283001 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2dmpc" Oct 10 13:31:20 crc kubenswrapper[4745]: I1010 13:31:20.284079 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qzh4v" event={"ID":"e6828db3-0a2c-4b98-8cbf-1eba04f13bc5","Type":"ContainerStarted","Data":"ff7a1e0fa83f8af62dd0855f6c8a0e398d4fad9d98e25ade2970c6d4496c3bba"} Oct 10 13:31:20 crc kubenswrapper[4745]: I1010 13:31:20.284285 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qzh4v" Oct 10 13:31:20 crc kubenswrapper[4745]: I1010 13:31:20.286636 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-xxvtr" event={"ID":"e1c2fd77-699a-4c79-9c54-9251c5ad7d17","Type":"ContainerStarted","Data":"4a20b1219bc09806731294f74317dc7d8651d6157cec657eb48d92df5d7df33c"} Oct 10 13:31:20 crc kubenswrapper[4745]: I1010 13:31:20.286790 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-xxvtr" Oct 10 13:31:20 crc kubenswrapper[4745]: I1010 13:31:20.288373 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-rz48f" event={"ID":"0ec7257b-b487-453c-88e8-d279184acdfc","Type":"ContainerStarted","Data":"5289669d3b84c55c59f5abf2ed554709e7d01afb1af73daab4e2a76694f832e6"} Oct 10 13:31:20 crc kubenswrapper[4745]: I1010 13:31:20.288573 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-rz48f" Oct 10 13:31:20 crc kubenswrapper[4745]: I1010 13:31:20.313978 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2dmpc" podStartSLOduration=2.920341704 podStartE2EDuration="27.31395936s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:55.567514002 +0000 UTC m=+769.465170765" lastFinishedPulling="2025-10-10 13:31:19.961131648 +0000 UTC m=+793.858788421" observedRunningTime="2025-10-10 13:31:20.298198527 +0000 UTC m=+794.195855300" watchObservedRunningTime="2025-10-10 13:31:20.31395936 +0000 UTC m=+794.211616133" Oct 10 13:31:20 crc kubenswrapper[4745]: I1010 13:31:20.314245 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-xxvtr" podStartSLOduration=3.042812255 podStartE2EDuration="27.314241287s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:55.658181041 +0000 UTC m=+769.555837804" lastFinishedPulling="2025-10-10 13:31:19.929610053 +0000 UTC m=+793.827266836" observedRunningTime="2025-10-10 13:31:20.312199815 +0000 UTC m=+794.209856588" watchObservedRunningTime="2025-10-10 13:31:20.314241287 +0000 UTC m=+794.211898040" Oct 10 13:31:20 crc kubenswrapper[4745]: I1010 13:31:20.331696 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-rz48f" podStartSLOduration=2.944322521 podStartE2EDuration="27.331678213s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:55.573040928 +0000 UTC m=+769.470697691" lastFinishedPulling="2025-10-10 13:31:19.96039659 +0000 UTC m=+793.858053383" observedRunningTime="2025-10-10 13:31:20.330186205 +0000 UTC m=+794.227842978" watchObservedRunningTime="2025-10-10 13:31:20.331678213 +0000 UTC m=+794.229334986" Oct 10 13:31:20 crc kubenswrapper[4745]: I1010 13:31:20.349653 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qzh4v" podStartSLOduration=3.144189657 podStartE2EDuration="27.349628662s" podCreationTimestamp="2025-10-10 13:30:53 +0000 UTC" firstStartedPulling="2025-10-10 13:30:55.749641079 +0000 UTC m=+769.647297842" lastFinishedPulling="2025-10-10 13:31:19.955080074 +0000 UTC m=+793.852736847" observedRunningTime="2025-10-10 13:31:20.345655181 +0000 UTC m=+794.243311964" watchObservedRunningTime="2025-10-10 13:31:20.349628662 +0000 UTC m=+794.247285425" Oct 10 13:31:20 crc kubenswrapper[4745]: I1010 13:31:20.752664 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15e38293-23c6-4705-9aaf-1de10eb3fe1e" path="/var/lib/kubelet/pods/15e38293-23c6-4705-9aaf-1de10eb3fe1e/volumes" Oct 10 13:31:24 crc kubenswrapper[4745]: I1010 13:31:24.229199 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-664664cb68-b7w5k" Oct 10 13:31:24 crc kubenswrapper[4745]: I1010 13:31:24.338415 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-lr987" Oct 10 13:31:24 crc kubenswrapper[4745]: I1010 13:31:24.395674 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-646675d848-96mn7" Oct 10 13:31:34 crc kubenswrapper[4745]: I1010 13:31:34.034478 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2dmpc" Oct 10 13:31:34 crc kubenswrapper[4745]: I1010 13:31:34.057779 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-xxvtr" Oct 10 13:31:34 crc kubenswrapper[4745]: I1010 13:31:34.197478 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-rz48f" Oct 10 13:31:34 crc kubenswrapper[4745]: I1010 13:31:34.298317 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-qzh4v" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.794866 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-96vgt"] Oct 10 13:31:50 crc kubenswrapper[4745]: E1010 13:31:50.795583 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e605d21-bc5f-44ee-a3f7-7558d307162b" containerName="registry-server" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.795596 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e605d21-bc5f-44ee-a3f7-7558d307162b" containerName="registry-server" Oct 10 13:31:50 crc kubenswrapper[4745]: E1010 13:31:50.795628 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15e38293-23c6-4705-9aaf-1de10eb3fe1e" containerName="registry-server" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.795635 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="15e38293-23c6-4705-9aaf-1de10eb3fe1e" containerName="registry-server" Oct 10 13:31:50 crc kubenswrapper[4745]: E1010 13:31:50.795648 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15e38293-23c6-4705-9aaf-1de10eb3fe1e" containerName="extract-content" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.795654 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="15e38293-23c6-4705-9aaf-1de10eb3fe1e" containerName="extract-content" Oct 10 13:31:50 crc kubenswrapper[4745]: E1010 13:31:50.795670 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15e38293-23c6-4705-9aaf-1de10eb3fe1e" containerName="extract-utilities" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.795675 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="15e38293-23c6-4705-9aaf-1de10eb3fe1e" containerName="extract-utilities" Oct 10 13:31:50 crc kubenswrapper[4745]: E1010 13:31:50.795682 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e605d21-bc5f-44ee-a3f7-7558d307162b" containerName="extract-utilities" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.795688 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e605d21-bc5f-44ee-a3f7-7558d307162b" containerName="extract-utilities" Oct 10 13:31:50 crc kubenswrapper[4745]: E1010 13:31:50.795699 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e605d21-bc5f-44ee-a3f7-7558d307162b" containerName="extract-content" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.795704 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e605d21-bc5f-44ee-a3f7-7558d307162b" containerName="extract-content" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.795886 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="15e38293-23c6-4705-9aaf-1de10eb3fe1e" containerName="registry-server" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.795908 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e605d21-bc5f-44ee-a3f7-7558d307162b" containerName="registry-server" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.796546 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-96vgt" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.802318 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.802326 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.802496 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-ztzq6" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.803875 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.811298 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-96vgt"] Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.957106 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-r4q4b"] Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.958818 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-r4q4b" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.960799 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.966544 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-r4q4b"] Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.987475 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2skwh\" (UniqueName: \"kubernetes.io/projected/f8f7b8ae-5b33-4944-9898-1c693fd3eb8d-kube-api-access-2skwh\") pod \"dnsmasq-dns-675f4bcbfc-96vgt\" (UID: \"f8f7b8ae-5b33-4944-9898-1c693fd3eb8d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-96vgt" Oct 10 13:31:50 crc kubenswrapper[4745]: I1010 13:31:50.987553 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8f7b8ae-5b33-4944-9898-1c693fd3eb8d-config\") pod \"dnsmasq-dns-675f4bcbfc-96vgt\" (UID: \"f8f7b8ae-5b33-4944-9898-1c693fd3eb8d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-96vgt" Oct 10 13:31:51 crc kubenswrapper[4745]: I1010 13:31:51.089312 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34b04641-f8b2-423b-870e-e15bd2b04a49-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-r4q4b\" (UID: \"34b04641-f8b2-423b-870e-e15bd2b04a49\") " pod="openstack/dnsmasq-dns-78dd6ddcc-r4q4b" Oct 10 13:31:51 crc kubenswrapper[4745]: I1010 13:31:51.089386 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8f7b8ae-5b33-4944-9898-1c693fd3eb8d-config\") pod \"dnsmasq-dns-675f4bcbfc-96vgt\" (UID: \"f8f7b8ae-5b33-4944-9898-1c693fd3eb8d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-96vgt" Oct 10 13:31:51 crc kubenswrapper[4745]: I1010 13:31:51.089411 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjb8b\" (UniqueName: \"kubernetes.io/projected/34b04641-f8b2-423b-870e-e15bd2b04a49-kube-api-access-mjb8b\") pod \"dnsmasq-dns-78dd6ddcc-r4q4b\" (UID: \"34b04641-f8b2-423b-870e-e15bd2b04a49\") " pod="openstack/dnsmasq-dns-78dd6ddcc-r4q4b" Oct 10 13:31:51 crc kubenswrapper[4745]: I1010 13:31:51.089492 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2skwh\" (UniqueName: \"kubernetes.io/projected/f8f7b8ae-5b33-4944-9898-1c693fd3eb8d-kube-api-access-2skwh\") pod \"dnsmasq-dns-675f4bcbfc-96vgt\" (UID: \"f8f7b8ae-5b33-4944-9898-1c693fd3eb8d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-96vgt" Oct 10 13:31:51 crc kubenswrapper[4745]: I1010 13:31:51.089520 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34b04641-f8b2-423b-870e-e15bd2b04a49-config\") pod \"dnsmasq-dns-78dd6ddcc-r4q4b\" (UID: \"34b04641-f8b2-423b-870e-e15bd2b04a49\") " pod="openstack/dnsmasq-dns-78dd6ddcc-r4q4b" Oct 10 13:31:51 crc kubenswrapper[4745]: I1010 13:31:51.090480 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8f7b8ae-5b33-4944-9898-1c693fd3eb8d-config\") pod \"dnsmasq-dns-675f4bcbfc-96vgt\" (UID: \"f8f7b8ae-5b33-4944-9898-1c693fd3eb8d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-96vgt" Oct 10 13:31:51 crc kubenswrapper[4745]: I1010 13:31:51.112598 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2skwh\" (UniqueName: \"kubernetes.io/projected/f8f7b8ae-5b33-4944-9898-1c693fd3eb8d-kube-api-access-2skwh\") pod \"dnsmasq-dns-675f4bcbfc-96vgt\" (UID: \"f8f7b8ae-5b33-4944-9898-1c693fd3eb8d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-96vgt" Oct 10 13:31:51 crc kubenswrapper[4745]: I1010 13:31:51.120761 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-96vgt" Oct 10 13:31:51 crc kubenswrapper[4745]: I1010 13:31:51.190884 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34b04641-f8b2-423b-870e-e15bd2b04a49-config\") pod \"dnsmasq-dns-78dd6ddcc-r4q4b\" (UID: \"34b04641-f8b2-423b-870e-e15bd2b04a49\") " pod="openstack/dnsmasq-dns-78dd6ddcc-r4q4b" Oct 10 13:31:51 crc kubenswrapper[4745]: I1010 13:31:51.190953 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34b04641-f8b2-423b-870e-e15bd2b04a49-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-r4q4b\" (UID: \"34b04641-f8b2-423b-870e-e15bd2b04a49\") " pod="openstack/dnsmasq-dns-78dd6ddcc-r4q4b" Oct 10 13:31:51 crc kubenswrapper[4745]: I1010 13:31:51.191008 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjb8b\" (UniqueName: \"kubernetes.io/projected/34b04641-f8b2-423b-870e-e15bd2b04a49-kube-api-access-mjb8b\") pod \"dnsmasq-dns-78dd6ddcc-r4q4b\" (UID: \"34b04641-f8b2-423b-870e-e15bd2b04a49\") " pod="openstack/dnsmasq-dns-78dd6ddcc-r4q4b" Oct 10 13:31:51 crc kubenswrapper[4745]: I1010 13:31:51.192262 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34b04641-f8b2-423b-870e-e15bd2b04a49-config\") pod \"dnsmasq-dns-78dd6ddcc-r4q4b\" (UID: \"34b04641-f8b2-423b-870e-e15bd2b04a49\") " pod="openstack/dnsmasq-dns-78dd6ddcc-r4q4b" Oct 10 13:31:51 crc kubenswrapper[4745]: I1010 13:31:51.192278 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34b04641-f8b2-423b-870e-e15bd2b04a49-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-r4q4b\" (UID: \"34b04641-f8b2-423b-870e-e15bd2b04a49\") " pod="openstack/dnsmasq-dns-78dd6ddcc-r4q4b" Oct 10 13:31:51 crc kubenswrapper[4745]: I1010 13:31:51.209920 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjb8b\" (UniqueName: \"kubernetes.io/projected/34b04641-f8b2-423b-870e-e15bd2b04a49-kube-api-access-mjb8b\") pod \"dnsmasq-dns-78dd6ddcc-r4q4b\" (UID: \"34b04641-f8b2-423b-870e-e15bd2b04a49\") " pod="openstack/dnsmasq-dns-78dd6ddcc-r4q4b" Oct 10 13:31:51 crc kubenswrapper[4745]: I1010 13:31:51.277219 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-r4q4b" Oct 10 13:31:51 crc kubenswrapper[4745]: I1010 13:31:51.591557 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-96vgt"] Oct 10 13:31:51 crc kubenswrapper[4745]: I1010 13:31:51.688552 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-r4q4b"] Oct 10 13:31:51 crc kubenswrapper[4745]: W1010 13:31:51.692299 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34b04641_f8b2_423b_870e_e15bd2b04a49.slice/crio-ac571230ddef41b07a1bec0d5f5f0d77c3ae55c1d08ff44ca8975aaa17efd48f WatchSource:0}: Error finding container ac571230ddef41b07a1bec0d5f5f0d77c3ae55c1d08ff44ca8975aaa17efd48f: Status 404 returned error can't find the container with id ac571230ddef41b07a1bec0d5f5f0d77c3ae55c1d08ff44ca8975aaa17efd48f Oct 10 13:31:52 crc kubenswrapper[4745]: I1010 13:31:52.551469 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-96vgt" event={"ID":"f8f7b8ae-5b33-4944-9898-1c693fd3eb8d","Type":"ContainerStarted","Data":"7a03254b4db7a1803f0d6d8d857986a45449e806f20cccef2b1420399b9cc22a"} Oct 10 13:31:52 crc kubenswrapper[4745]: I1010 13:31:52.554710 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-r4q4b" event={"ID":"34b04641-f8b2-423b-870e-e15bd2b04a49","Type":"ContainerStarted","Data":"ac571230ddef41b07a1bec0d5f5f0d77c3ae55c1d08ff44ca8975aaa17efd48f"} Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.453781 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-96vgt"] Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.460032 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-86nq5"] Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.462493 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-86nq5" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.469932 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-86nq5"] Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.552902 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69200f29-34cc-4b52-914b-25e8366dc5c7-dns-svc\") pod \"dnsmasq-dns-666b6646f7-86nq5\" (UID: \"69200f29-34cc-4b52-914b-25e8366dc5c7\") " pod="openstack/dnsmasq-dns-666b6646f7-86nq5" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.552958 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69200f29-34cc-4b52-914b-25e8366dc5c7-config\") pod \"dnsmasq-dns-666b6646f7-86nq5\" (UID: \"69200f29-34cc-4b52-914b-25e8366dc5c7\") " pod="openstack/dnsmasq-dns-666b6646f7-86nq5" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.553430 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jcld\" (UniqueName: \"kubernetes.io/projected/69200f29-34cc-4b52-914b-25e8366dc5c7-kube-api-access-9jcld\") pod \"dnsmasq-dns-666b6646f7-86nq5\" (UID: \"69200f29-34cc-4b52-914b-25e8366dc5c7\") " pod="openstack/dnsmasq-dns-666b6646f7-86nq5" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.654377 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jcld\" (UniqueName: \"kubernetes.io/projected/69200f29-34cc-4b52-914b-25e8366dc5c7-kube-api-access-9jcld\") pod \"dnsmasq-dns-666b6646f7-86nq5\" (UID: \"69200f29-34cc-4b52-914b-25e8366dc5c7\") " pod="openstack/dnsmasq-dns-666b6646f7-86nq5" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.654470 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69200f29-34cc-4b52-914b-25e8366dc5c7-dns-svc\") pod \"dnsmasq-dns-666b6646f7-86nq5\" (UID: \"69200f29-34cc-4b52-914b-25e8366dc5c7\") " pod="openstack/dnsmasq-dns-666b6646f7-86nq5" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.654505 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69200f29-34cc-4b52-914b-25e8366dc5c7-config\") pod \"dnsmasq-dns-666b6646f7-86nq5\" (UID: \"69200f29-34cc-4b52-914b-25e8366dc5c7\") " pod="openstack/dnsmasq-dns-666b6646f7-86nq5" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.655270 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69200f29-34cc-4b52-914b-25e8366dc5c7-dns-svc\") pod \"dnsmasq-dns-666b6646f7-86nq5\" (UID: \"69200f29-34cc-4b52-914b-25e8366dc5c7\") " pod="openstack/dnsmasq-dns-666b6646f7-86nq5" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.655345 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69200f29-34cc-4b52-914b-25e8366dc5c7-config\") pod \"dnsmasq-dns-666b6646f7-86nq5\" (UID: \"69200f29-34cc-4b52-914b-25e8366dc5c7\") " pod="openstack/dnsmasq-dns-666b6646f7-86nq5" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.681891 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jcld\" (UniqueName: \"kubernetes.io/projected/69200f29-34cc-4b52-914b-25e8366dc5c7-kube-api-access-9jcld\") pod \"dnsmasq-dns-666b6646f7-86nq5\" (UID: \"69200f29-34cc-4b52-914b-25e8366dc5c7\") " pod="openstack/dnsmasq-dns-666b6646f7-86nq5" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.737239 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-r4q4b"] Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.771094 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-59l6c"] Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.772217 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.785417 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-59l6c"] Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.792714 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-86nq5" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.857184 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgf9k\" (UniqueName: \"kubernetes.io/projected/ea2c879a-affe-4aab-9a97-037c1257bab7-kube-api-access-rgf9k\") pod \"dnsmasq-dns-57d769cc4f-59l6c\" (UID: \"ea2c879a-affe-4aab-9a97-037c1257bab7\") " pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.857263 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea2c879a-affe-4aab-9a97-037c1257bab7-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-59l6c\" (UID: \"ea2c879a-affe-4aab-9a97-037c1257bab7\") " pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.857302 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea2c879a-affe-4aab-9a97-037c1257bab7-config\") pod \"dnsmasq-dns-57d769cc4f-59l6c\" (UID: \"ea2c879a-affe-4aab-9a97-037c1257bab7\") " pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.958948 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgf9k\" (UniqueName: \"kubernetes.io/projected/ea2c879a-affe-4aab-9a97-037c1257bab7-kube-api-access-rgf9k\") pod \"dnsmasq-dns-57d769cc4f-59l6c\" (UID: \"ea2c879a-affe-4aab-9a97-037c1257bab7\") " pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.959019 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea2c879a-affe-4aab-9a97-037c1257bab7-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-59l6c\" (UID: \"ea2c879a-affe-4aab-9a97-037c1257bab7\") " pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.959059 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea2c879a-affe-4aab-9a97-037c1257bab7-config\") pod \"dnsmasq-dns-57d769cc4f-59l6c\" (UID: \"ea2c879a-affe-4aab-9a97-037c1257bab7\") " pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.959818 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea2c879a-affe-4aab-9a97-037c1257bab7-config\") pod \"dnsmasq-dns-57d769cc4f-59l6c\" (UID: \"ea2c879a-affe-4aab-9a97-037c1257bab7\") " pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.959872 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea2c879a-affe-4aab-9a97-037c1257bab7-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-59l6c\" (UID: \"ea2c879a-affe-4aab-9a97-037c1257bab7\") " pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" Oct 10 13:31:53 crc kubenswrapper[4745]: I1010 13:31:53.999125 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgf9k\" (UniqueName: \"kubernetes.io/projected/ea2c879a-affe-4aab-9a97-037c1257bab7-kube-api-access-rgf9k\") pod \"dnsmasq-dns-57d769cc4f-59l6c\" (UID: \"ea2c879a-affe-4aab-9a97-037c1257bab7\") " pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.091425 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.482633 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-86nq5"] Oct 10 13:31:54 crc kubenswrapper[4745]: W1010 13:31:54.489076 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69200f29_34cc_4b52_914b_25e8366dc5c7.slice/crio-6774e066842a99aa8094ca279885cdf7180065e22e5e30c4ff12abc8da44c123 WatchSource:0}: Error finding container 6774e066842a99aa8094ca279885cdf7180065e22e5e30c4ff12abc8da44c123: Status 404 returned error can't find the container with id 6774e066842a99aa8094ca279885cdf7180065e22e5e30c4ff12abc8da44c123 Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.574280 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-86nq5" event={"ID":"69200f29-34cc-4b52-914b-25e8366dc5c7","Type":"ContainerStarted","Data":"6774e066842a99aa8094ca279885cdf7180065e22e5e30c4ff12abc8da44c123"} Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.616542 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-59l6c"] Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.624892 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.625955 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.629219 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.629378 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.629517 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.629669 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.629821 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.629969 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-tkvfq" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.630080 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.639384 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.676241 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.676279 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5g5m\" (UniqueName: \"kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-kube-api-access-r5g5m\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.676302 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.676395 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.676453 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.676549 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.676593 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-config-data\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.676690 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.676771 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.676834 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.676882 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.782191 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.782540 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-config-data\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.782829 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.782895 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.782979 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.783021 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.783099 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.783119 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5g5m\" (UniqueName: \"kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-kube-api-access-r5g5m\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.783144 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.783185 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.783251 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.784818 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.787224 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.788470 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.788596 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.790618 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.792664 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.793264 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-config-data\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.793664 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.799339 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.799794 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.799995 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5g5m\" (UniqueName: \"kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-kube-api-access-r5g5m\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.815600 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.911601 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.912792 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.917161 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.917175 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.917513 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.917701 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.917940 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-hs5bv" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.917983 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.919206 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.928946 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.961923 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.986483 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/234155a2-9751-459f-b259-6bb32eb4b4d7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.986544 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.986927 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.986958 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.986978 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44b7l\" (UniqueName: \"kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-kube-api-access-44b7l\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.987411 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.987471 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.987653 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.987692 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.987715 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:54 crc kubenswrapper[4745]: I1010 13:31:54.987932 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/234155a2-9751-459f-b259-6bb32eb4b4d7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.089557 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.089624 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.089640 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.089668 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.090068 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/234155a2-9751-459f-b259-6bb32eb4b4d7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.090109 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/234155a2-9751-459f-b259-6bb32eb4b4d7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.090140 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.090155 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.090172 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.090482 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44b7l\" (UniqueName: \"kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-kube-api-access-44b7l\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.090500 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.090555 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.090672 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.091041 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.091159 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.091549 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.094466 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.093956 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/234155a2-9751-459f-b259-6bb32eb4b4d7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.094864 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/234155a2-9751-459f-b259-6bb32eb4b4d7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.095519 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.105882 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.107315 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44b7l\" (UniqueName: \"kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-kube-api-access-44b7l\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.118003 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.231170 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:31:55 crc kubenswrapper[4745]: I1010 13:31:55.584316 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" event={"ID":"ea2c879a-affe-4aab-9a97-037c1257bab7","Type":"ContainerStarted","Data":"4695bed8155b267ce2bfc530fa09cef29f80f2206c1cdc65affe2e7f9f1c523f"} Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.524640 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.529091 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.533341 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.533876 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.534088 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.534295 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-9wj9f" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.536764 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.540947 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.543283 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.611263 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35e75435-406e-4053-a1c2-2a7fc992b31e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.611350 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcmpz\" (UniqueName: \"kubernetes.io/projected/35e75435-406e-4053-a1c2-2a7fc992b31e-kube-api-access-tcmpz\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.611405 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.611471 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/35e75435-406e-4053-a1c2-2a7fc992b31e-config-data-default\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.611497 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/35e75435-406e-4053-a1c2-2a7fc992b31e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.611607 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/35e75435-406e-4053-a1c2-2a7fc992b31e-kolla-config\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.611649 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/35e75435-406e-4053-a1c2-2a7fc992b31e-secrets\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.611668 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e75435-406e-4053-a1c2-2a7fc992b31e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.611686 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/35e75435-406e-4053-a1c2-2a7fc992b31e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.713253 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35e75435-406e-4053-a1c2-2a7fc992b31e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.713316 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcmpz\" (UniqueName: \"kubernetes.io/projected/35e75435-406e-4053-a1c2-2a7fc992b31e-kube-api-access-tcmpz\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.713364 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.713394 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/35e75435-406e-4053-a1c2-2a7fc992b31e-config-data-default\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.713416 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/35e75435-406e-4053-a1c2-2a7fc992b31e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.713502 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/35e75435-406e-4053-a1c2-2a7fc992b31e-kolla-config\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.713581 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/35e75435-406e-4053-a1c2-2a7fc992b31e-secrets\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.713606 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e75435-406e-4053-a1c2-2a7fc992b31e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.713631 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/35e75435-406e-4053-a1c2-2a7fc992b31e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.713906 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.714206 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/35e75435-406e-4053-a1c2-2a7fc992b31e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.714609 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/35e75435-406e-4053-a1c2-2a7fc992b31e-config-data-default\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.715699 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/35e75435-406e-4053-a1c2-2a7fc992b31e-kolla-config\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.716358 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35e75435-406e-4053-a1c2-2a7fc992b31e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.718690 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e75435-406e-4053-a1c2-2a7fc992b31e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.719524 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/35e75435-406e-4053-a1c2-2a7fc992b31e-secrets\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.721378 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/35e75435-406e-4053-a1c2-2a7fc992b31e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.743348 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.750446 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcmpz\" (UniqueName: \"kubernetes.io/projected/35e75435-406e-4053-a1c2-2a7fc992b31e-kube-api-access-tcmpz\") pod \"openstack-galera-0\" (UID: \"35e75435-406e-4053-a1c2-2a7fc992b31e\") " pod="openstack/openstack-galera-0" Oct 10 13:31:56 crc kubenswrapper[4745]: I1010 13:31:56.908243 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.066507 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.069884 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.073280 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.073589 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.073598 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-gp7m2" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.074328 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.111541 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.238559 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c22070ab-5d89-42fa-b04a-640ec3045982-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.238619 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c22070ab-5d89-42fa-b04a-640ec3045982-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.238638 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c22070ab-5d89-42fa-b04a-640ec3045982-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.238677 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c22070ab-5d89-42fa-b04a-640ec3045982-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.238700 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c22070ab-5d89-42fa-b04a-640ec3045982-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.238750 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/c22070ab-5d89-42fa-b04a-640ec3045982-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.238780 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjrrg\" (UniqueName: \"kubernetes.io/projected/c22070ab-5d89-42fa-b04a-640ec3045982-kube-api-access-cjrrg\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.238802 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.238825 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c22070ab-5d89-42fa-b04a-640ec3045982-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.294133 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.295552 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.299263 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.299446 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-ms7zs" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.299565 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.304050 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.340633 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjrrg\" (UniqueName: \"kubernetes.io/projected/c22070ab-5d89-42fa-b04a-640ec3045982-kube-api-access-cjrrg\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.340685 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.340717 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c22070ab-5d89-42fa-b04a-640ec3045982-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.340930 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.341255 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c22070ab-5d89-42fa-b04a-640ec3045982-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.342639 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c22070ab-5d89-42fa-b04a-640ec3045982-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.342677 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c22070ab-5d89-42fa-b04a-640ec3045982-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.342716 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c22070ab-5d89-42fa-b04a-640ec3045982-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.342941 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c22070ab-5d89-42fa-b04a-640ec3045982-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.342986 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c22070ab-5d89-42fa-b04a-640ec3045982-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.343035 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/c22070ab-5d89-42fa-b04a-640ec3045982-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.343545 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c22070ab-5d89-42fa-b04a-640ec3045982-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.350937 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/c22070ab-5d89-42fa-b04a-640ec3045982-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.350988 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c22070ab-5d89-42fa-b04a-640ec3045982-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.352025 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c22070ab-5d89-42fa-b04a-640ec3045982-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.352661 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c22070ab-5d89-42fa-b04a-640ec3045982-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.359490 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c22070ab-5d89-42fa-b04a-640ec3045982-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.362269 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjrrg\" (UniqueName: \"kubernetes.io/projected/c22070ab-5d89-42fa-b04a-640ec3045982-kube-api-access-cjrrg\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.382856 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c22070ab-5d89-42fa-b04a-640ec3045982\") " pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.405652 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.443972 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e79afba-7f24-4e67-9c37-973256aac339-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9e79afba-7f24-4e67-9c37-973256aac339\") " pod="openstack/memcached-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.444023 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e79afba-7f24-4e67-9c37-973256aac339-config-data\") pod \"memcached-0\" (UID: \"9e79afba-7f24-4e67-9c37-973256aac339\") " pod="openstack/memcached-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.444044 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e79afba-7f24-4e67-9c37-973256aac339-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9e79afba-7f24-4e67-9c37-973256aac339\") " pod="openstack/memcached-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.444061 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9e79afba-7f24-4e67-9c37-973256aac339-kolla-config\") pod \"memcached-0\" (UID: \"9e79afba-7f24-4e67-9c37-973256aac339\") " pod="openstack/memcached-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.444132 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5p8x\" (UniqueName: \"kubernetes.io/projected/9e79afba-7f24-4e67-9c37-973256aac339-kube-api-access-g5p8x\") pod \"memcached-0\" (UID: \"9e79afba-7f24-4e67-9c37-973256aac339\") " pod="openstack/memcached-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.545301 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5p8x\" (UniqueName: \"kubernetes.io/projected/9e79afba-7f24-4e67-9c37-973256aac339-kube-api-access-g5p8x\") pod \"memcached-0\" (UID: \"9e79afba-7f24-4e67-9c37-973256aac339\") " pod="openstack/memcached-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.545371 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e79afba-7f24-4e67-9c37-973256aac339-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9e79afba-7f24-4e67-9c37-973256aac339\") " pod="openstack/memcached-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.545406 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e79afba-7f24-4e67-9c37-973256aac339-config-data\") pod \"memcached-0\" (UID: \"9e79afba-7f24-4e67-9c37-973256aac339\") " pod="openstack/memcached-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.545433 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e79afba-7f24-4e67-9c37-973256aac339-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9e79afba-7f24-4e67-9c37-973256aac339\") " pod="openstack/memcached-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.545453 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9e79afba-7f24-4e67-9c37-973256aac339-kolla-config\") pod \"memcached-0\" (UID: \"9e79afba-7f24-4e67-9c37-973256aac339\") " pod="openstack/memcached-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.546290 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9e79afba-7f24-4e67-9c37-973256aac339-kolla-config\") pod \"memcached-0\" (UID: \"9e79afba-7f24-4e67-9c37-973256aac339\") " pod="openstack/memcached-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.546782 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e79afba-7f24-4e67-9c37-973256aac339-config-data\") pod \"memcached-0\" (UID: \"9e79afba-7f24-4e67-9c37-973256aac339\") " pod="openstack/memcached-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.549119 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e79afba-7f24-4e67-9c37-973256aac339-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9e79afba-7f24-4e67-9c37-973256aac339\") " pod="openstack/memcached-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.550020 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e79afba-7f24-4e67-9c37-973256aac339-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9e79afba-7f24-4e67-9c37-973256aac339\") " pod="openstack/memcached-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.565823 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5p8x\" (UniqueName: \"kubernetes.io/projected/9e79afba-7f24-4e67-9c37-973256aac339-kube-api-access-g5p8x\") pod \"memcached-0\" (UID: \"9e79afba-7f24-4e67-9c37-973256aac339\") " pod="openstack/memcached-0" Oct 10 13:31:58 crc kubenswrapper[4745]: I1010 13:31:58.623426 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 10 13:31:59 crc kubenswrapper[4745]: I1010 13:31:59.978184 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 13:31:59 crc kubenswrapper[4745]: I1010 13:31:59.979347 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 13:31:59 crc kubenswrapper[4745]: I1010 13:31:59.981616 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-hxstb" Oct 10 13:31:59 crc kubenswrapper[4745]: I1010 13:31:59.986028 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 13:32:00 crc kubenswrapper[4745]: I1010 13:32:00.166591 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4pgr\" (UniqueName: \"kubernetes.io/projected/2f59de38-158a-4c6d-b902-daed00239a27-kube-api-access-b4pgr\") pod \"kube-state-metrics-0\" (UID: \"2f59de38-158a-4c6d-b902-daed00239a27\") " pod="openstack/kube-state-metrics-0" Oct 10 13:32:00 crc kubenswrapper[4745]: I1010 13:32:00.267782 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4pgr\" (UniqueName: \"kubernetes.io/projected/2f59de38-158a-4c6d-b902-daed00239a27-kube-api-access-b4pgr\") pod \"kube-state-metrics-0\" (UID: \"2f59de38-158a-4c6d-b902-daed00239a27\") " pod="openstack/kube-state-metrics-0" Oct 10 13:32:00 crc kubenswrapper[4745]: I1010 13:32:00.293491 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4pgr\" (UniqueName: \"kubernetes.io/projected/2f59de38-158a-4c6d-b902-daed00239a27-kube-api-access-b4pgr\") pod \"kube-state-metrics-0\" (UID: \"2f59de38-158a-4c6d-b902-daed00239a27\") " pod="openstack/kube-state-metrics-0" Oct 10 13:32:00 crc kubenswrapper[4745]: I1010 13:32:00.297959 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 13:32:03 crc kubenswrapper[4745]: I1010 13:32:03.043325 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x77sv"] Oct 10 13:32:03 crc kubenswrapper[4745]: I1010 13:32:03.045336 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x77sv" Oct 10 13:32:03 crc kubenswrapper[4745]: I1010 13:32:03.053048 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x77sv"] Oct 10 13:32:03 crc kubenswrapper[4745]: I1010 13:32:03.222573 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335ddca3-4773-4797-85b2-49b959c06fc3-utilities\") pod \"community-operators-x77sv\" (UID: \"335ddca3-4773-4797-85b2-49b959c06fc3\") " pod="openshift-marketplace/community-operators-x77sv" Oct 10 13:32:03 crc kubenswrapper[4745]: I1010 13:32:03.222639 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335ddca3-4773-4797-85b2-49b959c06fc3-catalog-content\") pod \"community-operators-x77sv\" (UID: \"335ddca3-4773-4797-85b2-49b959c06fc3\") " pod="openshift-marketplace/community-operators-x77sv" Oct 10 13:32:03 crc kubenswrapper[4745]: I1010 13:32:03.222680 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkzrv\" (UniqueName: \"kubernetes.io/projected/335ddca3-4773-4797-85b2-49b959c06fc3-kube-api-access-vkzrv\") pod \"community-operators-x77sv\" (UID: \"335ddca3-4773-4797-85b2-49b959c06fc3\") " pod="openshift-marketplace/community-operators-x77sv" Oct 10 13:32:03 crc kubenswrapper[4745]: I1010 13:32:03.326520 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkzrv\" (UniqueName: \"kubernetes.io/projected/335ddca3-4773-4797-85b2-49b959c06fc3-kube-api-access-vkzrv\") pod \"community-operators-x77sv\" (UID: \"335ddca3-4773-4797-85b2-49b959c06fc3\") " pod="openshift-marketplace/community-operators-x77sv" Oct 10 13:32:03 crc kubenswrapper[4745]: I1010 13:32:03.326621 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335ddca3-4773-4797-85b2-49b959c06fc3-utilities\") pod \"community-operators-x77sv\" (UID: \"335ddca3-4773-4797-85b2-49b959c06fc3\") " pod="openshift-marketplace/community-operators-x77sv" Oct 10 13:32:03 crc kubenswrapper[4745]: I1010 13:32:03.326668 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335ddca3-4773-4797-85b2-49b959c06fc3-catalog-content\") pod \"community-operators-x77sv\" (UID: \"335ddca3-4773-4797-85b2-49b959c06fc3\") " pod="openshift-marketplace/community-operators-x77sv" Oct 10 13:32:03 crc kubenswrapper[4745]: I1010 13:32:03.327254 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335ddca3-4773-4797-85b2-49b959c06fc3-catalog-content\") pod \"community-operators-x77sv\" (UID: \"335ddca3-4773-4797-85b2-49b959c06fc3\") " pod="openshift-marketplace/community-operators-x77sv" Oct 10 13:32:03 crc kubenswrapper[4745]: I1010 13:32:03.327445 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335ddca3-4773-4797-85b2-49b959c06fc3-utilities\") pod \"community-operators-x77sv\" (UID: \"335ddca3-4773-4797-85b2-49b959c06fc3\") " pod="openshift-marketplace/community-operators-x77sv" Oct 10 13:32:03 crc kubenswrapper[4745]: I1010 13:32:03.348672 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkzrv\" (UniqueName: \"kubernetes.io/projected/335ddca3-4773-4797-85b2-49b959c06fc3-kube-api-access-vkzrv\") pod \"community-operators-x77sv\" (UID: \"335ddca3-4773-4797-85b2-49b959c06fc3\") " pod="openshift-marketplace/community-operators-x77sv" Oct 10 13:32:03 crc kubenswrapper[4745]: I1010 13:32:03.376264 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x77sv" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.026227 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-92pgn"] Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.032888 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.034643 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.035132 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.035456 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-jmx8l" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.043381 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-92pgn"] Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.047512 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-s2qdr"] Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.048892 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.096755 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-s2qdr"] Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.140384 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513dcce5-f68e-42a4-ba94-44013577e96e-combined-ca-bundle\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.140433 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/513dcce5-f68e-42a4-ba94-44013577e96e-var-log-ovn\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.140479 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e9b37d55-6160-43d6-8957-4cc31c95251d-var-lib\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.140506 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e9b37d55-6160-43d6-8957-4cc31c95251d-var-log\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.140528 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/513dcce5-f68e-42a4-ba94-44013577e96e-scripts\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.140544 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/513dcce5-f68e-42a4-ba94-44013577e96e-ovn-controller-tls-certs\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.140572 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/513dcce5-f68e-42a4-ba94-44013577e96e-var-run\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.140596 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/513dcce5-f68e-42a4-ba94-44013577e96e-var-run-ovn\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.140628 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9b37d55-6160-43d6-8957-4cc31c95251d-scripts\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.140647 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74prg\" (UniqueName: \"kubernetes.io/projected/e9b37d55-6160-43d6-8957-4cc31c95251d-kube-api-access-74prg\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.140664 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e9b37d55-6160-43d6-8957-4cc31c95251d-etc-ovs\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.140684 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e9b37d55-6160-43d6-8957-4cc31c95251d-var-run\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.140700 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wgwp\" (UniqueName: \"kubernetes.io/projected/513dcce5-f68e-42a4-ba94-44013577e96e-kube-api-access-5wgwp\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.242190 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e9b37d55-6160-43d6-8957-4cc31c95251d-var-lib\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.242243 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e9b37d55-6160-43d6-8957-4cc31c95251d-var-log\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.242267 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/513dcce5-f68e-42a4-ba94-44013577e96e-scripts\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.242284 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/513dcce5-f68e-42a4-ba94-44013577e96e-ovn-controller-tls-certs\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.242314 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/513dcce5-f68e-42a4-ba94-44013577e96e-var-run\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.242338 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/513dcce5-f68e-42a4-ba94-44013577e96e-var-run-ovn\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.244292 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/513dcce5-f68e-42a4-ba94-44013577e96e-var-run-ovn\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.244343 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/513dcce5-f68e-42a4-ba94-44013577e96e-var-run\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.244401 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9b37d55-6160-43d6-8957-4cc31c95251d-scripts\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.244392 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e9b37d55-6160-43d6-8957-4cc31c95251d-var-log\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.244446 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74prg\" (UniqueName: \"kubernetes.io/projected/e9b37d55-6160-43d6-8957-4cc31c95251d-kube-api-access-74prg\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.244445 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e9b37d55-6160-43d6-8957-4cc31c95251d-var-lib\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.244474 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e9b37d55-6160-43d6-8957-4cc31c95251d-etc-ovs\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.244566 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e9b37d55-6160-43d6-8957-4cc31c95251d-var-run\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.244610 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wgwp\" (UniqueName: \"kubernetes.io/projected/513dcce5-f68e-42a4-ba94-44013577e96e-kube-api-access-5wgwp\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.244655 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e9b37d55-6160-43d6-8957-4cc31c95251d-var-run\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.244688 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513dcce5-f68e-42a4-ba94-44013577e96e-combined-ca-bundle\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.244956 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/513dcce5-f68e-42a4-ba94-44013577e96e-var-log-ovn\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.244613 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e9b37d55-6160-43d6-8957-4cc31c95251d-etc-ovs\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.246013 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/513dcce5-f68e-42a4-ba94-44013577e96e-var-log-ovn\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.246012 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/513dcce5-f68e-42a4-ba94-44013577e96e-scripts\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.247146 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9b37d55-6160-43d6-8957-4cc31c95251d-scripts\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.250390 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513dcce5-f68e-42a4-ba94-44013577e96e-combined-ca-bundle\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.271442 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/513dcce5-f68e-42a4-ba94-44013577e96e-ovn-controller-tls-certs\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.274674 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wgwp\" (UniqueName: \"kubernetes.io/projected/513dcce5-f68e-42a4-ba94-44013577e96e-kube-api-access-5wgwp\") pod \"ovn-controller-92pgn\" (UID: \"513dcce5-f68e-42a4-ba94-44013577e96e\") " pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.283395 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74prg\" (UniqueName: \"kubernetes.io/projected/e9b37d55-6160-43d6-8957-4cc31c95251d-kube-api-access-74prg\") pod \"ovn-controller-ovs-s2qdr\" (UID: \"e9b37d55-6160-43d6-8957-4cc31c95251d\") " pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.408396 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-92pgn" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.431818 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.903197 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.904773 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.908514 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.908915 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.909017 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.909235 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-bsd2k" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.909594 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.918712 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.956878 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd428c33-2518-41ce-8dd4-b62fe3059525-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.956940 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.956965 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dd428c33-2518-41ce-8dd4-b62fe3059525-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.957036 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd428c33-2518-41ce-8dd4-b62fe3059525-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.957057 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd428c33-2518-41ce-8dd4-b62fe3059525-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.957090 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd428c33-2518-41ce-8dd4-b62fe3059525-config\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.957152 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dgfd\" (UniqueName: \"kubernetes.io/projected/dd428c33-2518-41ce-8dd4-b62fe3059525-kube-api-access-4dgfd\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:04 crc kubenswrapper[4745]: I1010 13:32:04.957191 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd428c33-2518-41ce-8dd4-b62fe3059525-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.058605 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd428c33-2518-41ce-8dd4-b62fe3059525-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.058651 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd428c33-2518-41ce-8dd4-b62fe3059525-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.058691 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd428c33-2518-41ce-8dd4-b62fe3059525-config\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.058761 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dgfd\" (UniqueName: \"kubernetes.io/projected/dd428c33-2518-41ce-8dd4-b62fe3059525-kube-api-access-4dgfd\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.058805 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd428c33-2518-41ce-8dd4-b62fe3059525-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.058864 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd428c33-2518-41ce-8dd4-b62fe3059525-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.058897 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.058921 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dd428c33-2518-41ce-8dd4-b62fe3059525-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.059491 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dd428c33-2518-41ce-8dd4-b62fe3059525-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.059807 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.060040 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd428c33-2518-41ce-8dd4-b62fe3059525-config\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.060522 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd428c33-2518-41ce-8dd4-b62fe3059525-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.063535 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd428c33-2518-41ce-8dd4-b62fe3059525-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.066959 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd428c33-2518-41ce-8dd4-b62fe3059525-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.071634 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd428c33-2518-41ce-8dd4-b62fe3059525-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.077349 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dgfd\" (UniqueName: \"kubernetes.io/projected/dd428c33-2518-41ce-8dd4-b62fe3059525-kube-api-access-4dgfd\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.084066 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"dd428c33-2518-41ce-8dd4-b62fe3059525\") " pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:05 crc kubenswrapper[4745]: I1010 13:32:05.234085 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:06 crc kubenswrapper[4745]: I1010 13:32:06.289556 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 13:32:06 crc kubenswrapper[4745]: E1010 13:32:06.706138 4745 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 10 13:32:06 crc kubenswrapper[4745]: E1010 13:32:06.706295 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mjb8b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-r4q4b_openstack(34b04641-f8b2-423b-870e-e15bd2b04a49): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 13:32:06 crc kubenswrapper[4745]: E1010 13:32:06.708325 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-r4q4b" podUID="34b04641-f8b2-423b-870e-e15bd2b04a49" Oct 10 13:32:06 crc kubenswrapper[4745]: E1010 13:32:06.794439 4745 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 10 13:32:06 crc kubenswrapper[4745]: E1010 13:32:06.794607 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2skwh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-96vgt_openstack(f8f7b8ae-5b33-4944-9898-1c693fd3eb8d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 13:32:06 crc kubenswrapper[4745]: E1010 13:32:06.796124 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-96vgt" podUID="f8f7b8ae-5b33-4944-9898-1c693fd3eb8d" Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.124712 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 10 13:32:07 crc kubenswrapper[4745]: W1010 13:32:07.186714 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35e75435_406e_4053_a1c2_2a7fc992b31e.slice/crio-a4278a430fd7641adb6b099a2afca880f42cde424e3d035741e8843fd6cd63df WatchSource:0}: Error finding container a4278a430fd7641adb6b099a2afca880f42cde424e3d035741e8843fd6cd63df: Status 404 returned error can't find the container with id a4278a430fd7641adb6b099a2afca880f42cde424e3d035741e8843fd6cd63df Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.300778 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 10 13:32:07 crc kubenswrapper[4745]: W1010 13:32:07.302109 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e79afba_7f24_4e67_9c37_973256aac339.slice/crio-772e79e11e32f7344f943ab51b73cdf5a04647f2eec1c7637800d9135c170ae1 WatchSource:0}: Error finding container 772e79e11e32f7344f943ab51b73cdf5a04647f2eec1c7637800d9135c170ae1: Status 404 returned error can't find the container with id 772e79e11e32f7344f943ab51b73cdf5a04647f2eec1c7637800d9135c170ae1 Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.388659 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.486283 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.499637 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.507916 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-92pgn"] Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.515025 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x77sv"] Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.521275 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 13:32:07 crc kubenswrapper[4745]: W1010 13:32:07.537353 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f59de38_158a_4c6d_b902_daed00239a27.slice/crio-cccdd845bf25820d8e10ffd92184d358a1b2c2603d53ab7ae17dca89bf6a6f26 WatchSource:0}: Error finding container cccdd845bf25820d8e10ffd92184d358a1b2c2603d53ab7ae17dca89bf6a6f26: Status 404 returned error can't find the container with id cccdd845bf25820d8e10ffd92184d358a1b2c2603d53ab7ae17dca89bf6a6f26 Oct 10 13:32:07 crc kubenswrapper[4745]: W1010 13:32:07.539235 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod234155a2_9751_459f_b259_6bb32eb4b4d7.slice/crio-3b8c5a01627d4585f00f719d1bd453a1eac0f2e8bae0eeabd5c4e20f8ff2eaae WatchSource:0}: Error finding container 3b8c5a01627d4585f00f719d1bd453a1eac0f2e8bae0eeabd5c4e20f8ff2eaae: Status 404 returned error can't find the container with id 3b8c5a01627d4585f00f719d1bd453a1eac0f2e8bae0eeabd5c4e20f8ff2eaae Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.665877 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-s2qdr"] Oct 10 13:32:07 crc kubenswrapper[4745]: W1010 13:32:07.676846 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9b37d55_6160_43d6_8957_4cc31c95251d.slice/crio-1485ecb5697a2e0331d469bb50706796933de50d01324ca1986325a3826fff47 WatchSource:0}: Error finding container 1485ecb5697a2e0331d469bb50706796933de50d01324ca1986325a3826fff47: Status 404 returned error can't find the container with id 1485ecb5697a2e0331d469bb50706796933de50d01324ca1986325a3826fff47 Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.690231 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c22070ab-5d89-42fa-b04a-640ec3045982","Type":"ContainerStarted","Data":"54f0531d42c4479a70a7d73d77274bde81f268f4ab8c76e100edd338a49f47d7"} Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.691380 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bc0bb94e-36a7-413d-9ce0-eea2031d1d19","Type":"ContainerStarted","Data":"26e4f12fcb056d774ac4b8b7f3ca3349a8a2a9165229875488fd6158fc9fdfe6"} Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.692628 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"234155a2-9751-459f-b259-6bb32eb4b4d7","Type":"ContainerStarted","Data":"3b8c5a01627d4585f00f719d1bd453a1eac0f2e8bae0eeabd5c4e20f8ff2eaae"} Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.694292 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-92pgn" event={"ID":"513dcce5-f68e-42a4-ba94-44013577e96e","Type":"ContainerStarted","Data":"bcb5bb7909b659961029b86a0bec4c10e5a23cd06cd2d5b8723f31a8f28189d0"} Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.696025 4745 generic.go:334] "Generic (PLEG): container finished" podID="ea2c879a-affe-4aab-9a97-037c1257bab7" containerID="17b5cedd3d31763693ba178e917c7fab4d454e1811c9125a0043e73f9c2de4df" exitCode=0 Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.696201 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" event={"ID":"ea2c879a-affe-4aab-9a97-037c1257bab7","Type":"ContainerDied","Data":"17b5cedd3d31763693ba178e917c7fab4d454e1811c9125a0043e73f9c2de4df"} Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.699261 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2f59de38-158a-4c6d-b902-daed00239a27","Type":"ContainerStarted","Data":"cccdd845bf25820d8e10ffd92184d358a1b2c2603d53ab7ae17dca89bf6a6f26"} Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.700352 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9e79afba-7f24-4e67-9c37-973256aac339","Type":"ContainerStarted","Data":"772e79e11e32f7344f943ab51b73cdf5a04647f2eec1c7637800d9135c170ae1"} Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.703237 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-s2qdr" event={"ID":"e9b37d55-6160-43d6-8957-4cc31c95251d","Type":"ContainerStarted","Data":"1485ecb5697a2e0331d469bb50706796933de50d01324ca1986325a3826fff47"} Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.704873 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"35e75435-406e-4053-a1c2-2a7fc992b31e","Type":"ContainerStarted","Data":"a4278a430fd7641adb6b099a2afca880f42cde424e3d035741e8843fd6cd63df"} Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.706250 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"dd428c33-2518-41ce-8dd4-b62fe3059525","Type":"ContainerStarted","Data":"e2bc82b1a44338931ab91d902c338a7298e3ea8097598f145b901f04b7244057"} Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.708404 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x77sv" event={"ID":"335ddca3-4773-4797-85b2-49b959c06fc3","Type":"ContainerStarted","Data":"66ca45c8e132a9c4f12b94e1888e1f77f15cdfef94962b1b59dceac5eeb319ee"} Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.708440 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x77sv" event={"ID":"335ddca3-4773-4797-85b2-49b959c06fc3","Type":"ContainerStarted","Data":"ae566cd25a2985e83a37d16c011128e28fb61914353f1806469a7364a7210b59"} Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.714974 4745 generic.go:334] "Generic (PLEG): container finished" podID="69200f29-34cc-4b52-914b-25e8366dc5c7" containerID="111e6b9056b5ac8ec2899a05ef8d07e716227873a3139cc5bd8819bdb11a3030" exitCode=0 Oct 10 13:32:07 crc kubenswrapper[4745]: I1010 13:32:07.715046 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-86nq5" event={"ID":"69200f29-34cc-4b52-914b-25e8366dc5c7","Type":"ContainerDied","Data":"111e6b9056b5ac8ec2899a05ef8d07e716227873a3139cc5bd8819bdb11a3030"} Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.093955 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.096562 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.098832 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.098991 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.099122 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.099229 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-tjz78" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.101171 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.156183 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-r4q4b" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.162688 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-96vgt" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.216180 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3894eba1-5402-4b7f-9859-a57827e29f2b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.216223 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.216253 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3894eba1-5402-4b7f-9859-a57827e29f2b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.216270 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3894eba1-5402-4b7f-9859-a57827e29f2b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.216295 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3894eba1-5402-4b7f-9859-a57827e29f2b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.216320 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbrls\" (UniqueName: \"kubernetes.io/projected/3894eba1-5402-4b7f-9859-a57827e29f2b-kube-api-access-bbrls\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.216353 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3894eba1-5402-4b7f-9859-a57827e29f2b-config\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.216400 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3894eba1-5402-4b7f-9859-a57827e29f2b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.317853 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2skwh\" (UniqueName: \"kubernetes.io/projected/f8f7b8ae-5b33-4944-9898-1c693fd3eb8d-kube-api-access-2skwh\") pod \"f8f7b8ae-5b33-4944-9898-1c693fd3eb8d\" (UID: \"f8f7b8ae-5b33-4944-9898-1c693fd3eb8d\") " Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.318102 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34b04641-f8b2-423b-870e-e15bd2b04a49-config\") pod \"34b04641-f8b2-423b-870e-e15bd2b04a49\" (UID: \"34b04641-f8b2-423b-870e-e15bd2b04a49\") " Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.318139 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34b04641-f8b2-423b-870e-e15bd2b04a49-dns-svc\") pod \"34b04641-f8b2-423b-870e-e15bd2b04a49\" (UID: \"34b04641-f8b2-423b-870e-e15bd2b04a49\") " Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.318196 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8f7b8ae-5b33-4944-9898-1c693fd3eb8d-config\") pod \"f8f7b8ae-5b33-4944-9898-1c693fd3eb8d\" (UID: \"f8f7b8ae-5b33-4944-9898-1c693fd3eb8d\") " Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.318215 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjb8b\" (UniqueName: \"kubernetes.io/projected/34b04641-f8b2-423b-870e-e15bd2b04a49-kube-api-access-mjb8b\") pod \"34b04641-f8b2-423b-870e-e15bd2b04a49\" (UID: \"34b04641-f8b2-423b-870e-e15bd2b04a49\") " Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.318459 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3894eba1-5402-4b7f-9859-a57827e29f2b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.318514 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3894eba1-5402-4b7f-9859-a57827e29f2b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.318541 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.318565 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3894eba1-5402-4b7f-9859-a57827e29f2b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.318580 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3894eba1-5402-4b7f-9859-a57827e29f2b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.318611 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3894eba1-5402-4b7f-9859-a57827e29f2b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.318638 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbrls\" (UniqueName: \"kubernetes.io/projected/3894eba1-5402-4b7f-9859-a57827e29f2b-kube-api-access-bbrls\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.318674 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3894eba1-5402-4b7f-9859-a57827e29f2b-config\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.320934 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.320953 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8f7b8ae-5b33-4944-9898-1c693fd3eb8d-config" (OuterVolumeSpecName: "config") pod "f8f7b8ae-5b33-4944-9898-1c693fd3eb8d" (UID: "f8f7b8ae-5b33-4944-9898-1c693fd3eb8d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.321020 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34b04641-f8b2-423b-870e-e15bd2b04a49-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "34b04641-f8b2-423b-870e-e15bd2b04a49" (UID: "34b04641-f8b2-423b-870e-e15bd2b04a49"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.321187 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3894eba1-5402-4b7f-9859-a57827e29f2b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.321554 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34b04641-f8b2-423b-870e-e15bd2b04a49-config" (OuterVolumeSpecName: "config") pod "34b04641-f8b2-423b-870e-e15bd2b04a49" (UID: "34b04641-f8b2-423b-870e-e15bd2b04a49"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.321744 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3894eba1-5402-4b7f-9859-a57827e29f2b-config\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.324480 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34b04641-f8b2-423b-870e-e15bd2b04a49-kube-api-access-mjb8b" (OuterVolumeSpecName: "kube-api-access-mjb8b") pod "34b04641-f8b2-423b-870e-e15bd2b04a49" (UID: "34b04641-f8b2-423b-870e-e15bd2b04a49"). InnerVolumeSpecName "kube-api-access-mjb8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.325630 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3894eba1-5402-4b7f-9859-a57827e29f2b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.327375 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3894eba1-5402-4b7f-9859-a57827e29f2b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.327559 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3894eba1-5402-4b7f-9859-a57827e29f2b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.327596 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8f7b8ae-5b33-4944-9898-1c693fd3eb8d-kube-api-access-2skwh" (OuterVolumeSpecName: "kube-api-access-2skwh") pod "f8f7b8ae-5b33-4944-9898-1c693fd3eb8d" (UID: "f8f7b8ae-5b33-4944-9898-1c693fd3eb8d"). InnerVolumeSpecName "kube-api-access-2skwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.331540 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3894eba1-5402-4b7f-9859-a57827e29f2b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.337177 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbrls\" (UniqueName: \"kubernetes.io/projected/3894eba1-5402-4b7f-9859-a57827e29f2b-kube-api-access-bbrls\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.347825 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"3894eba1-5402-4b7f-9859-a57827e29f2b\") " pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.421603 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2skwh\" (UniqueName: \"kubernetes.io/projected/f8f7b8ae-5b33-4944-9898-1c693fd3eb8d-kube-api-access-2skwh\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.422328 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34b04641-f8b2-423b-870e-e15bd2b04a49-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.423083 4745 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34b04641-f8b2-423b-870e-e15bd2b04a49-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.423131 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8f7b8ae-5b33-4944-9898-1c693fd3eb8d-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.423146 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjb8b\" (UniqueName: \"kubernetes.io/projected/34b04641-f8b2-423b-870e-e15bd2b04a49-kube-api-access-mjb8b\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.427553 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.726111 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-96vgt" event={"ID":"f8f7b8ae-5b33-4944-9898-1c693fd3eb8d","Type":"ContainerDied","Data":"7a03254b4db7a1803f0d6d8d857986a45449e806f20cccef2b1420399b9cc22a"} Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.726303 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-96vgt" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.731793 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" event={"ID":"ea2c879a-affe-4aab-9a97-037c1257bab7","Type":"ContainerStarted","Data":"37420428e1b96a7a14f7025b5067ba0dd81a01d75572df3f35a8257a3b7d3685"} Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.732009 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.736706 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-86nq5" event={"ID":"69200f29-34cc-4b52-914b-25e8366dc5c7","Type":"ContainerStarted","Data":"e512791388b5924ff16ccb0e53606e6412e7dda8edf2cc65e98e42566803ad26"} Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.737560 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-86nq5" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.739270 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-r4q4b" event={"ID":"34b04641-f8b2-423b-870e-e15bd2b04a49","Type":"ContainerDied","Data":"ac571230ddef41b07a1bec0d5f5f0d77c3ae55c1d08ff44ca8975aaa17efd48f"} Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.739330 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-r4q4b" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.743155 4745 generic.go:334] "Generic (PLEG): container finished" podID="335ddca3-4773-4797-85b2-49b959c06fc3" containerID="66ca45c8e132a9c4f12b94e1888e1f77f15cdfef94962b1b59dceac5eeb319ee" exitCode=0 Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.743199 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x77sv" event={"ID":"335ddca3-4773-4797-85b2-49b959c06fc3","Type":"ContainerDied","Data":"66ca45c8e132a9c4f12b94e1888e1f77f15cdfef94962b1b59dceac5eeb319ee"} Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.758703 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" podStartSLOduration=3.592618119 podStartE2EDuration="15.75867944s" podCreationTimestamp="2025-10-10 13:31:53 +0000 UTC" firstStartedPulling="2025-10-10 13:31:54.638185746 +0000 UTC m=+828.535842499" lastFinishedPulling="2025-10-10 13:32:06.804247057 +0000 UTC m=+840.701903820" observedRunningTime="2025-10-10 13:32:08.752386769 +0000 UTC m=+842.650043532" watchObservedRunningTime="2025-10-10 13:32:08.75867944 +0000 UTC m=+842.656336203" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.785750 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-86nq5" podStartSLOduration=3.475005062 podStartE2EDuration="15.785716431s" podCreationTimestamp="2025-10-10 13:31:53 +0000 UTC" firstStartedPulling="2025-10-10 13:31:54.491988588 +0000 UTC m=+828.389645351" lastFinishedPulling="2025-10-10 13:32:06.802699957 +0000 UTC m=+840.700356720" observedRunningTime="2025-10-10 13:32:08.770017249 +0000 UTC m=+842.667674013" watchObservedRunningTime="2025-10-10 13:32:08.785716431 +0000 UTC m=+842.683373194" Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.812011 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-96vgt"] Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.817983 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-96vgt"] Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.833209 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-r4q4b"] Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.838131 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-r4q4b"] Oct 10 13:32:08 crc kubenswrapper[4745]: I1010 13:32:08.928065 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.315918 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-xhz46"] Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.316827 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.319699 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.334155 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-xhz46"] Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.451526 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-config\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.451576 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.451878 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-ovs-rundir\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.451908 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-combined-ca-bundle\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.452040 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf2lf\" (UniqueName: \"kubernetes.io/projected/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-kube-api-access-mf2lf\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.452065 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-ovn-rundir\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.468257 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-59l6c"] Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.490869 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-cm4r8"] Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.496399 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.498251 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.509260 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-cm4r8"] Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.553076 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf2lf\" (UniqueName: \"kubernetes.io/projected/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-kube-api-access-mf2lf\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.553119 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-ovn-rundir\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.553170 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-config\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.553190 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.553239 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-ovs-rundir\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.553266 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-combined-ca-bundle\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.554542 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-config\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.554943 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-ovn-rundir\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.557648 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-ovs-rundir\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.559467 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.559681 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-combined-ca-bundle\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.582356 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf2lf\" (UniqueName: \"kubernetes.io/projected/3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce-kube-api-access-mf2lf\") pod \"ovn-controller-metrics-xhz46\" (UID: \"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce\") " pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.607712 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-86nq5"] Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.628995 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r2kc2"] Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.637766 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.639359 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.645347 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r2kc2"] Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.657621 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22rzr\" (UniqueName: \"kubernetes.io/projected/5d9cada8-cb95-44d1-846b-0b415548ef15-kube-api-access-22rzr\") pod \"dnsmasq-dns-7fd796d7df-cm4r8\" (UID: \"5d9cada8-cb95-44d1-846b-0b415548ef15\") " pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.657668 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-cm4r8\" (UID: \"5d9cada8-cb95-44d1-846b-0b415548ef15\") " pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.657767 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-config\") pod \"dnsmasq-dns-7fd796d7df-cm4r8\" (UID: \"5d9cada8-cb95-44d1-846b-0b415548ef15\") " pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.657798 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-cm4r8\" (UID: \"5d9cada8-cb95-44d1-846b-0b415548ef15\") " pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.662351 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-xhz46" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.759826 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-config\") pod \"dnsmasq-dns-7fd796d7df-cm4r8\" (UID: \"5d9cada8-cb95-44d1-846b-0b415548ef15\") " pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.759875 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-r2kc2\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.759919 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-cm4r8\" (UID: \"5d9cada8-cb95-44d1-846b-0b415548ef15\") " pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.760101 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22rzr\" (UniqueName: \"kubernetes.io/projected/5d9cada8-cb95-44d1-846b-0b415548ef15-kube-api-access-22rzr\") pod \"dnsmasq-dns-7fd796d7df-cm4r8\" (UID: \"5d9cada8-cb95-44d1-846b-0b415548ef15\") " pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.760151 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-cm4r8\" (UID: \"5d9cada8-cb95-44d1-846b-0b415548ef15\") " pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.760191 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm7x5\" (UniqueName: \"kubernetes.io/projected/75ea909b-a461-4053-8a51-6b804e365c4d-kube-api-access-bm7x5\") pod \"dnsmasq-dns-86db49b7ff-r2kc2\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.760212 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-r2kc2\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.760242 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-r2kc2\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.760311 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-config\") pod \"dnsmasq-dns-86db49b7ff-r2kc2\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.760912 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-config\") pod \"dnsmasq-dns-7fd796d7df-cm4r8\" (UID: \"5d9cada8-cb95-44d1-846b-0b415548ef15\") " pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.761649 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-cm4r8\" (UID: \"5d9cada8-cb95-44d1-846b-0b415548ef15\") " pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.765325 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-cm4r8\" (UID: \"5d9cada8-cb95-44d1-846b-0b415548ef15\") " pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.778427 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22rzr\" (UniqueName: \"kubernetes.io/projected/5d9cada8-cb95-44d1-846b-0b415548ef15-kube-api-access-22rzr\") pod \"dnsmasq-dns-7fd796d7df-cm4r8\" (UID: \"5d9cada8-cb95-44d1-846b-0b415548ef15\") " pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.831559 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.861547 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-config\") pod \"dnsmasq-dns-86db49b7ff-r2kc2\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.861652 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-r2kc2\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.861769 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm7x5\" (UniqueName: \"kubernetes.io/projected/75ea909b-a461-4053-8a51-6b804e365c4d-kube-api-access-bm7x5\") pod \"dnsmasq-dns-86db49b7ff-r2kc2\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.861790 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-r2kc2\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.861821 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-r2kc2\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.862651 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-r2kc2\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.863809 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-r2kc2\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.863986 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-r2kc2\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.864365 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-config\") pod \"dnsmasq-dns-86db49b7ff-r2kc2\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.877673 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm7x5\" (UniqueName: \"kubernetes.io/projected/75ea909b-a461-4053-8a51-6b804e365c4d-kube-api-access-bm7x5\") pod \"dnsmasq-dns-86db49b7ff-r2kc2\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:09 crc kubenswrapper[4745]: I1010 13:32:09.980798 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:10 crc kubenswrapper[4745]: W1010 13:32:10.638375 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3894eba1_5402_4b7f_9859_a57827e29f2b.slice/crio-343bb5d4c4231451747915ebf51d0fc2d6c85e3989278578d89a2006d8d7c655 WatchSource:0}: Error finding container 343bb5d4c4231451747915ebf51d0fc2d6c85e3989278578d89a2006d8d7c655: Status 404 returned error can't find the container with id 343bb5d4c4231451747915ebf51d0fc2d6c85e3989278578d89a2006d8d7c655 Oct 10 13:32:10 crc kubenswrapper[4745]: I1010 13:32:10.758068 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34b04641-f8b2-423b-870e-e15bd2b04a49" path="/var/lib/kubelet/pods/34b04641-f8b2-423b-870e-e15bd2b04a49/volumes" Oct 10 13:32:10 crc kubenswrapper[4745]: I1010 13:32:10.758820 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8f7b8ae-5b33-4944-9898-1c693fd3eb8d" path="/var/lib/kubelet/pods/f8f7b8ae-5b33-4944-9898-1c693fd3eb8d/volumes" Oct 10 13:32:10 crc kubenswrapper[4745]: I1010 13:32:10.759948 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"3894eba1-5402-4b7f-9859-a57827e29f2b","Type":"ContainerStarted","Data":"343bb5d4c4231451747915ebf51d0fc2d6c85e3989278578d89a2006d8d7c655"} Oct 10 13:32:10 crc kubenswrapper[4745]: I1010 13:32:10.760086 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-86nq5" podUID="69200f29-34cc-4b52-914b-25e8366dc5c7" containerName="dnsmasq-dns" containerID="cri-o://e512791388b5924ff16ccb0e53606e6412e7dda8edf2cc65e98e42566803ad26" gracePeriod=10 Oct 10 13:32:10 crc kubenswrapper[4745]: I1010 13:32:10.760211 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" podUID="ea2c879a-affe-4aab-9a97-037c1257bab7" containerName="dnsmasq-dns" containerID="cri-o://37420428e1b96a7a14f7025b5067ba0dd81a01d75572df3f35a8257a3b7d3685" gracePeriod=10 Oct 10 13:32:11 crc kubenswrapper[4745]: I1010 13:32:11.769720 4745 generic.go:334] "Generic (PLEG): container finished" podID="ea2c879a-affe-4aab-9a97-037c1257bab7" containerID="37420428e1b96a7a14f7025b5067ba0dd81a01d75572df3f35a8257a3b7d3685" exitCode=0 Oct 10 13:32:11 crc kubenswrapper[4745]: I1010 13:32:11.769786 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" event={"ID":"ea2c879a-affe-4aab-9a97-037c1257bab7","Type":"ContainerDied","Data":"37420428e1b96a7a14f7025b5067ba0dd81a01d75572df3f35a8257a3b7d3685"} Oct 10 13:32:11 crc kubenswrapper[4745]: I1010 13:32:11.772169 4745 generic.go:334] "Generic (PLEG): container finished" podID="69200f29-34cc-4b52-914b-25e8366dc5c7" containerID="e512791388b5924ff16ccb0e53606e6412e7dda8edf2cc65e98e42566803ad26" exitCode=0 Oct 10 13:32:11 crc kubenswrapper[4745]: I1010 13:32:11.772205 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-86nq5" event={"ID":"69200f29-34cc-4b52-914b-25e8366dc5c7","Type":"ContainerDied","Data":"e512791388b5924ff16ccb0e53606e6412e7dda8edf2cc65e98e42566803ad26"} Oct 10 13:32:14 crc kubenswrapper[4745]: I1010 13:32:14.798406 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" Oct 10 13:32:14 crc kubenswrapper[4745]: I1010 13:32:14.802803 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" event={"ID":"ea2c879a-affe-4aab-9a97-037c1257bab7","Type":"ContainerDied","Data":"4695bed8155b267ce2bfc530fa09cef29f80f2206c1cdc65affe2e7f9f1c523f"} Oct 10 13:32:14 crc kubenswrapper[4745]: I1010 13:32:14.802971 4745 scope.go:117] "RemoveContainer" containerID="37420428e1b96a7a14f7025b5067ba0dd81a01d75572df3f35a8257a3b7d3685" Oct 10 13:32:14 crc kubenswrapper[4745]: I1010 13:32:14.802844 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" Oct 10 13:32:14 crc kubenswrapper[4745]: I1010 13:32:14.852657 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgf9k\" (UniqueName: \"kubernetes.io/projected/ea2c879a-affe-4aab-9a97-037c1257bab7-kube-api-access-rgf9k\") pod \"ea2c879a-affe-4aab-9a97-037c1257bab7\" (UID: \"ea2c879a-affe-4aab-9a97-037c1257bab7\") " Oct 10 13:32:14 crc kubenswrapper[4745]: I1010 13:32:14.852706 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea2c879a-affe-4aab-9a97-037c1257bab7-dns-svc\") pod \"ea2c879a-affe-4aab-9a97-037c1257bab7\" (UID: \"ea2c879a-affe-4aab-9a97-037c1257bab7\") " Oct 10 13:32:14 crc kubenswrapper[4745]: I1010 13:32:14.852745 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea2c879a-affe-4aab-9a97-037c1257bab7-config\") pod \"ea2c879a-affe-4aab-9a97-037c1257bab7\" (UID: \"ea2c879a-affe-4aab-9a97-037c1257bab7\") " Oct 10 13:32:14 crc kubenswrapper[4745]: I1010 13:32:14.863980 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea2c879a-affe-4aab-9a97-037c1257bab7-kube-api-access-rgf9k" (OuterVolumeSpecName: "kube-api-access-rgf9k") pod "ea2c879a-affe-4aab-9a97-037c1257bab7" (UID: "ea2c879a-affe-4aab-9a97-037c1257bab7"). InnerVolumeSpecName "kube-api-access-rgf9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:32:14 crc kubenswrapper[4745]: I1010 13:32:14.889538 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea2c879a-affe-4aab-9a97-037c1257bab7-config" (OuterVolumeSpecName: "config") pod "ea2c879a-affe-4aab-9a97-037c1257bab7" (UID: "ea2c879a-affe-4aab-9a97-037c1257bab7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:14 crc kubenswrapper[4745]: I1010 13:32:14.892629 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea2c879a-affe-4aab-9a97-037c1257bab7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ea2c879a-affe-4aab-9a97-037c1257bab7" (UID: "ea2c879a-affe-4aab-9a97-037c1257bab7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:14 crc kubenswrapper[4745]: I1010 13:32:14.953986 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgf9k\" (UniqueName: \"kubernetes.io/projected/ea2c879a-affe-4aab-9a97-037c1257bab7-kube-api-access-rgf9k\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:14 crc kubenswrapper[4745]: I1010 13:32:14.954020 4745 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea2c879a-affe-4aab-9a97-037c1257bab7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:14 crc kubenswrapper[4745]: I1010 13:32:14.954031 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea2c879a-affe-4aab-9a97-037c1257bab7-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:15 crc kubenswrapper[4745]: I1010 13:32:15.147496 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-59l6c"] Oct 10 13:32:15 crc kubenswrapper[4745]: I1010 13:32:15.154477 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-59l6c"] Oct 10 13:32:15 crc kubenswrapper[4745]: I1010 13:32:15.346076 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-86nq5" Oct 10 13:32:15 crc kubenswrapper[4745]: I1010 13:32:15.364784 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69200f29-34cc-4b52-914b-25e8366dc5c7-dns-svc\") pod \"69200f29-34cc-4b52-914b-25e8366dc5c7\" (UID: \"69200f29-34cc-4b52-914b-25e8366dc5c7\") " Oct 10 13:32:15 crc kubenswrapper[4745]: I1010 13:32:15.410336 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69200f29-34cc-4b52-914b-25e8366dc5c7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "69200f29-34cc-4b52-914b-25e8366dc5c7" (UID: "69200f29-34cc-4b52-914b-25e8366dc5c7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:15 crc kubenswrapper[4745]: I1010 13:32:15.466175 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69200f29-34cc-4b52-914b-25e8366dc5c7-config\") pod \"69200f29-34cc-4b52-914b-25e8366dc5c7\" (UID: \"69200f29-34cc-4b52-914b-25e8366dc5c7\") " Oct 10 13:32:15 crc kubenswrapper[4745]: I1010 13:32:15.466300 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jcld\" (UniqueName: \"kubernetes.io/projected/69200f29-34cc-4b52-914b-25e8366dc5c7-kube-api-access-9jcld\") pod \"69200f29-34cc-4b52-914b-25e8366dc5c7\" (UID: \"69200f29-34cc-4b52-914b-25e8366dc5c7\") " Oct 10 13:32:15 crc kubenswrapper[4745]: I1010 13:32:15.466481 4745 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69200f29-34cc-4b52-914b-25e8366dc5c7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:15 crc kubenswrapper[4745]: I1010 13:32:15.469034 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69200f29-34cc-4b52-914b-25e8366dc5c7-kube-api-access-9jcld" (OuterVolumeSpecName: "kube-api-access-9jcld") pod "69200f29-34cc-4b52-914b-25e8366dc5c7" (UID: "69200f29-34cc-4b52-914b-25e8366dc5c7"). InnerVolumeSpecName "kube-api-access-9jcld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:32:15 crc kubenswrapper[4745]: I1010 13:32:15.495261 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69200f29-34cc-4b52-914b-25e8366dc5c7-config" (OuterVolumeSpecName: "config") pod "69200f29-34cc-4b52-914b-25e8366dc5c7" (UID: "69200f29-34cc-4b52-914b-25e8366dc5c7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:15 crc kubenswrapper[4745]: I1010 13:32:15.567812 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jcld\" (UniqueName: \"kubernetes.io/projected/69200f29-34cc-4b52-914b-25e8366dc5c7-kube-api-access-9jcld\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:15 crc kubenswrapper[4745]: I1010 13:32:15.567850 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69200f29-34cc-4b52-914b-25e8366dc5c7-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:15 crc kubenswrapper[4745]: I1010 13:32:15.731785 4745 scope.go:117] "RemoveContainer" containerID="17b5cedd3d31763693ba178e917c7fab4d454e1811c9125a0043e73f9c2de4df" Oct 10 13:32:15 crc kubenswrapper[4745]: I1010 13:32:15.843981 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-86nq5" Oct 10 13:32:15 crc kubenswrapper[4745]: I1010 13:32:15.843949 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-86nq5" event={"ID":"69200f29-34cc-4b52-914b-25e8366dc5c7","Type":"ContainerDied","Data":"6774e066842a99aa8094ca279885cdf7180065e22e5e30c4ff12abc8da44c123"} Oct 10 13:32:15 crc kubenswrapper[4745]: I1010 13:32:15.883636 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-86nq5"] Oct 10 13:32:15 crc kubenswrapper[4745]: I1010 13:32:15.890192 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-86nq5"] Oct 10 13:32:16 crc kubenswrapper[4745]: I1010 13:32:16.177815 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-xhz46"] Oct 10 13:32:16 crc kubenswrapper[4745]: I1010 13:32:16.185369 4745 scope.go:117] "RemoveContainer" containerID="e512791388b5924ff16ccb0e53606e6412e7dda8edf2cc65e98e42566803ad26" Oct 10 13:32:16 crc kubenswrapper[4745]: W1010 13:32:16.190510 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f7ecb0e_ebe5_48f5_a89e_adf1f9fa01ce.slice/crio-1353973ff822ada9db184b451f8b49ca77214aaa8c818311640ea7602e0c7610 WatchSource:0}: Error finding container 1353973ff822ada9db184b451f8b49ca77214aaa8c818311640ea7602e0c7610: Status 404 returned error can't find the container with id 1353973ff822ada9db184b451f8b49ca77214aaa8c818311640ea7602e0c7610 Oct 10 13:32:16 crc kubenswrapper[4745]: I1010 13:32:16.195408 4745 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 13:32:16 crc kubenswrapper[4745]: I1010 13:32:16.299383 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-cm4r8"] Oct 10 13:32:16 crc kubenswrapper[4745]: I1010 13:32:16.363197 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r2kc2"] Oct 10 13:32:16 crc kubenswrapper[4745]: I1010 13:32:16.653276 4745 scope.go:117] "RemoveContainer" containerID="111e6b9056b5ac8ec2899a05ef8d07e716227873a3139cc5bd8819bdb11a3030" Oct 10 13:32:16 crc kubenswrapper[4745]: W1010 13:32:16.670300 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d9cada8_cb95_44d1_846b_0b415548ef15.slice/crio-00fc0998a0b80a514320703f3f2353d518a6df07138c534f71e68bfab1278235 WatchSource:0}: Error finding container 00fc0998a0b80a514320703f3f2353d518a6df07138c534f71e68bfab1278235: Status 404 returned error can't find the container with id 00fc0998a0b80a514320703f3f2353d518a6df07138c534f71e68bfab1278235 Oct 10 13:32:16 crc kubenswrapper[4745]: W1010 13:32:16.674580 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75ea909b_a461_4053_8a51_6b804e365c4d.slice/crio-3105a7463afc56b15d4ea8c91a0590f26dc5eebd5c62374030318a13044f4fa6 WatchSource:0}: Error finding container 3105a7463afc56b15d4ea8c91a0590f26dc5eebd5c62374030318a13044f4fa6: Status 404 returned error can't find the container with id 3105a7463afc56b15d4ea8c91a0590f26dc5eebd5c62374030318a13044f4fa6 Oct 10 13:32:16 crc kubenswrapper[4745]: I1010 13:32:16.779425 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69200f29-34cc-4b52-914b-25e8366dc5c7" path="/var/lib/kubelet/pods/69200f29-34cc-4b52-914b-25e8366dc5c7/volumes" Oct 10 13:32:16 crc kubenswrapper[4745]: I1010 13:32:16.781827 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea2c879a-affe-4aab-9a97-037c1257bab7" path="/var/lib/kubelet/pods/ea2c879a-affe-4aab-9a97-037c1257bab7/volumes" Oct 10 13:32:16 crc kubenswrapper[4745]: I1010 13:32:16.866941 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-xhz46" event={"ID":"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce","Type":"ContainerStarted","Data":"1353973ff822ada9db184b451f8b49ca77214aaa8c818311640ea7602e0c7610"} Oct 10 13:32:16 crc kubenswrapper[4745]: I1010 13:32:16.874506 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" event={"ID":"75ea909b-a461-4053-8a51-6b804e365c4d","Type":"ContainerStarted","Data":"3105a7463afc56b15d4ea8c91a0590f26dc5eebd5c62374030318a13044f4fa6"} Oct 10 13:32:16 crc kubenswrapper[4745]: I1010 13:32:16.899522 4745 generic.go:334] "Generic (PLEG): container finished" podID="335ddca3-4773-4797-85b2-49b959c06fc3" containerID="b705fe2b47c8b0fc110b6996f2f3e0912d7ae653a64cfd5144261cbc5eb8e342" exitCode=0 Oct 10 13:32:16 crc kubenswrapper[4745]: I1010 13:32:16.899932 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x77sv" event={"ID":"335ddca3-4773-4797-85b2-49b959c06fc3","Type":"ContainerDied","Data":"b705fe2b47c8b0fc110b6996f2f3e0912d7ae653a64cfd5144261cbc5eb8e342"} Oct 10 13:32:16 crc kubenswrapper[4745]: I1010 13:32:16.904269 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" event={"ID":"5d9cada8-cb95-44d1-846b-0b415548ef15","Type":"ContainerStarted","Data":"00fc0998a0b80a514320703f3f2353d518a6df07138c534f71e68bfab1278235"} Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.949851 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-92pgn" event={"ID":"513dcce5-f68e-42a4-ba94-44013577e96e","Type":"ContainerStarted","Data":"50f24257781e2269a1600e0932becd65604423039fc16cf86f99a6d7579ab913"} Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.950351 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-92pgn" Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.952633 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2f59de38-158a-4c6d-b902-daed00239a27","Type":"ContainerStarted","Data":"7220f11f38ff2c790ce1a10f2004b3a7b290380db4339fcad096fc5bbe639915"} Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.952884 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.957302 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9e79afba-7f24-4e67-9c37-973256aac339","Type":"ContainerStarted","Data":"8e57b9264eef49929bd3fe8a562b628fe524ff899ae44c41951e00331c09ed7e"} Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.958541 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.961425 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-s2qdr" event={"ID":"e9b37d55-6160-43d6-8957-4cc31c95251d","Type":"ContainerStarted","Data":"59d394292807be377f653c7578e9c2274a52b9c396fabcfc23e05d7631a3bd03"} Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.969868 4745 generic.go:334] "Generic (PLEG): container finished" podID="75ea909b-a461-4053-8a51-6b804e365c4d" containerID="d6915a05208f1c296ad397dff3c90c1ff6198ed7451ce9319fd013a76c6dc1c8" exitCode=0 Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.969926 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" event={"ID":"75ea909b-a461-4053-8a51-6b804e365c4d","Type":"ContainerDied","Data":"d6915a05208f1c296ad397dff3c90c1ff6198ed7451ce9319fd013a76c6dc1c8"} Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.971328 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"3894eba1-5402-4b7f-9859-a57827e29f2b","Type":"ContainerStarted","Data":"d0d69b859a0131fe301c4c1215096f527d510808266c6601a186c540e0b159ae"} Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.973129 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-92pgn" podStartSLOduration=5.314205419 podStartE2EDuration="13.973118564s" podCreationTimestamp="2025-10-10 13:32:04 +0000 UTC" firstStartedPulling="2025-10-10 13:32:07.526506534 +0000 UTC m=+841.424163297" lastFinishedPulling="2025-10-10 13:32:16.185419679 +0000 UTC m=+850.083076442" observedRunningTime="2025-10-10 13:32:17.963641585 +0000 UTC m=+851.861298348" watchObservedRunningTime="2025-10-10 13:32:17.973118564 +0000 UTC m=+851.870775327" Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.973621 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x77sv" event={"ID":"335ddca3-4773-4797-85b2-49b959c06fc3","Type":"ContainerStarted","Data":"0391359040ba5c79e7934d752a64c379fb7178bdc17249455733648e28b682d1"} Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.979236 4745 generic.go:334] "Generic (PLEG): container finished" podID="5d9cada8-cb95-44d1-846b-0b415548ef15" containerID="a29ab365ae90f38c4f1b24ec0ca6da87542866533e565cfa9258d73766deb283" exitCode=0 Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.979317 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" event={"ID":"5d9cada8-cb95-44d1-846b-0b415548ef15","Type":"ContainerDied","Data":"a29ab365ae90f38c4f1b24ec0ca6da87542866533e565cfa9258d73766deb283"} Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.982467 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"dd428c33-2518-41ce-8dd4-b62fe3059525","Type":"ContainerStarted","Data":"cf98f52223f587e0cfc340499e1eef45669277f76c70571c2712458376340257"} Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.983334 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=12.196076282 podStartE2EDuration="19.983315321s" podCreationTimestamp="2025-10-10 13:31:58 +0000 UTC" firstStartedPulling="2025-10-10 13:32:07.304106878 +0000 UTC m=+841.201763641" lastFinishedPulling="2025-10-10 13:32:15.091345917 +0000 UTC m=+848.989002680" observedRunningTime="2025-10-10 13:32:17.980009118 +0000 UTC m=+851.877665881" watchObservedRunningTime="2025-10-10 13:32:17.983315321 +0000 UTC m=+851.880972084" Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.984834 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"35e75435-406e-4053-a1c2-2a7fc992b31e","Type":"ContainerStarted","Data":"cecd616ad534be94ec45bcd6c52547014483ddcb1660c40a0991ae6173e0226d"} Oct 10 13:32:17 crc kubenswrapper[4745]: I1010 13:32:17.987047 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c22070ab-5d89-42fa-b04a-640ec3045982","Type":"ContainerStarted","Data":"f39f56bc4a0aabbd34d77ba5a81689b809b8aafe94c061ceb4bb4e6c4dded7ad"} Oct 10 13:32:18 crc kubenswrapper[4745]: I1010 13:32:18.015048 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=9.674211208 podStartE2EDuration="19.015029821s" podCreationTimestamp="2025-10-10 13:31:59 +0000 UTC" firstStartedPulling="2025-10-10 13:32:07.551146384 +0000 UTC m=+841.448803147" lastFinishedPulling="2025-10-10 13:32:16.891964997 +0000 UTC m=+850.789621760" observedRunningTime="2025-10-10 13:32:18.007954702 +0000 UTC m=+851.905611465" watchObservedRunningTime="2025-10-10 13:32:18.015029821 +0000 UTC m=+851.912686584" Oct 10 13:32:18 crc kubenswrapper[4745]: I1010 13:32:18.015834 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"234155a2-9751-459f-b259-6bb32eb4b4d7","Type":"ContainerStarted","Data":"ba8c59cb21a1b92f873d2755789e9b4ca0999c1be1d4d2bdf990e767e13ffec7"} Oct 10 13:32:18 crc kubenswrapper[4745]: I1010 13:32:18.035197 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x77sv" podStartSLOduration=5.315794564 podStartE2EDuration="15.035180729s" podCreationTimestamp="2025-10-10 13:32:03 +0000 UTC" firstStartedPulling="2025-10-10 13:32:07.709694678 +0000 UTC m=+841.607351441" lastFinishedPulling="2025-10-10 13:32:17.429080843 +0000 UTC m=+851.326737606" observedRunningTime="2025-10-10 13:32:18.028191893 +0000 UTC m=+851.925848656" watchObservedRunningTime="2025-10-10 13:32:18.035180729 +0000 UTC m=+851.932837492" Oct 10 13:32:18 crc kubenswrapper[4745]: I1010 13:32:18.794328 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-666b6646f7-86nq5" podUID="69200f29-34cc-4b52-914b-25e8366dc5c7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.97:5353: i/o timeout" Oct 10 13:32:19 crc kubenswrapper[4745]: I1010 13:32:19.027445 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bc0bb94e-36a7-413d-9ce0-eea2031d1d19","Type":"ContainerStarted","Data":"e9b2c0742bac99a38637cce65f207a0590e5ea6ba54dc0c7bcb8a39f75955347"} Oct 10 13:32:19 crc kubenswrapper[4745]: I1010 13:32:19.030568 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" event={"ID":"5d9cada8-cb95-44d1-846b-0b415548ef15","Type":"ContainerStarted","Data":"16eac1ccf9dbce2c4bc9030d0e822e406f75268d5ca2bbe3222daace4a4eab25"} Oct 10 13:32:19 crc kubenswrapper[4745]: I1010 13:32:19.030768 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:19 crc kubenswrapper[4745]: I1010 13:32:19.033205 4745 generic.go:334] "Generic (PLEG): container finished" podID="e9b37d55-6160-43d6-8957-4cc31c95251d" containerID="59d394292807be377f653c7578e9c2274a52b9c396fabcfc23e05d7631a3bd03" exitCode=0 Oct 10 13:32:19 crc kubenswrapper[4745]: I1010 13:32:19.033235 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-s2qdr" event={"ID":"e9b37d55-6160-43d6-8957-4cc31c95251d","Type":"ContainerDied","Data":"59d394292807be377f653c7578e9c2274a52b9c396fabcfc23e05d7631a3bd03"} Oct 10 13:32:19 crc kubenswrapper[4745]: I1010 13:32:19.033281 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-s2qdr" event={"ID":"e9b37d55-6160-43d6-8957-4cc31c95251d","Type":"ContainerStarted","Data":"51fd1a490cde925a29ee4b3763102b663ced4447b12960d7adc4e58460240868"} Oct 10 13:32:19 crc kubenswrapper[4745]: I1010 13:32:19.033296 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-s2qdr" event={"ID":"e9b37d55-6160-43d6-8957-4cc31c95251d","Type":"ContainerStarted","Data":"983babd9f1dec441896e737d21c1ea75d711cbe24ee0029569d68d9612722d8f"} Oct 10 13:32:19 crc kubenswrapper[4745]: I1010 13:32:19.033357 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:19 crc kubenswrapper[4745]: I1010 13:32:19.034980 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" event={"ID":"75ea909b-a461-4053-8a51-6b804e365c4d","Type":"ContainerStarted","Data":"c8726d44b9c7eeaa21f920a1b1fff755304f278c3a170417ea69f1bd0c361b67"} Oct 10 13:32:19 crc kubenswrapper[4745]: I1010 13:32:19.035851 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:19 crc kubenswrapper[4745]: I1010 13:32:19.074338 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" podStartSLOduration=10.074322426 podStartE2EDuration="10.074322426s" podCreationTimestamp="2025-10-10 13:32:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:32:19.07089446 +0000 UTC m=+852.968551213" watchObservedRunningTime="2025-10-10 13:32:19.074322426 +0000 UTC m=+852.971979179" Oct 10 13:32:19 crc kubenswrapper[4745]: I1010 13:32:19.088839 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-s2qdr" podStartSLOduration=7.138823309 podStartE2EDuration="15.088824892s" podCreationTimestamp="2025-10-10 13:32:04 +0000 UTC" firstStartedPulling="2025-10-10 13:32:07.678712016 +0000 UTC m=+841.576368779" lastFinishedPulling="2025-10-10 13:32:15.628713599 +0000 UTC m=+849.526370362" observedRunningTime="2025-10-10 13:32:19.086158615 +0000 UTC m=+852.983815378" watchObservedRunningTime="2025-10-10 13:32:19.088824892 +0000 UTC m=+852.986481645" Oct 10 13:32:19 crc kubenswrapper[4745]: I1010 13:32:19.092863 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57d769cc4f-59l6c" podUID="ea2c879a-affe-4aab-9a97-037c1257bab7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.98:5353: i/o timeout" Oct 10 13:32:19 crc kubenswrapper[4745]: I1010 13:32:19.109890 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" podStartSLOduration=10.109872163 podStartE2EDuration="10.109872163s" podCreationTimestamp="2025-10-10 13:32:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:32:19.105983475 +0000 UTC m=+853.003640238" watchObservedRunningTime="2025-10-10 13:32:19.109872163 +0000 UTC m=+853.007528926" Oct 10 13:32:19 crc kubenswrapper[4745]: I1010 13:32:19.432666 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:22 crc kubenswrapper[4745]: I1010 13:32:22.068627 4745 generic.go:334] "Generic (PLEG): container finished" podID="35e75435-406e-4053-a1c2-2a7fc992b31e" containerID="cecd616ad534be94ec45bcd6c52547014483ddcb1660c40a0991ae6173e0226d" exitCode=0 Oct 10 13:32:22 crc kubenswrapper[4745]: I1010 13:32:22.068762 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"35e75435-406e-4053-a1c2-2a7fc992b31e","Type":"ContainerDied","Data":"cecd616ad534be94ec45bcd6c52547014483ddcb1660c40a0991ae6173e0226d"} Oct 10 13:32:22 crc kubenswrapper[4745]: I1010 13:32:22.072013 4745 generic.go:334] "Generic (PLEG): container finished" podID="c22070ab-5d89-42fa-b04a-640ec3045982" containerID="f39f56bc4a0aabbd34d77ba5a81689b809b8aafe94c061ceb4bb4e6c4dded7ad" exitCode=0 Oct 10 13:32:22 crc kubenswrapper[4745]: I1010 13:32:22.072077 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c22070ab-5d89-42fa-b04a-640ec3045982","Type":"ContainerDied","Data":"f39f56bc4a0aabbd34d77ba5a81689b809b8aafe94c061ceb4bb4e6c4dded7ad"} Oct 10 13:32:23 crc kubenswrapper[4745]: I1010 13:32:23.377362 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x77sv" Oct 10 13:32:23 crc kubenswrapper[4745]: I1010 13:32:23.377404 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x77sv" Oct 10 13:32:23 crc kubenswrapper[4745]: I1010 13:32:23.455519 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x77sv" Oct 10 13:32:23 crc kubenswrapper[4745]: I1010 13:32:23.625333 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 10 13:32:24 crc kubenswrapper[4745]: I1010 13:32:24.163714 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x77sv" Oct 10 13:32:24 crc kubenswrapper[4745]: I1010 13:32:24.230490 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x77sv"] Oct 10 13:32:24 crc kubenswrapper[4745]: I1010 13:32:24.833874 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:24 crc kubenswrapper[4745]: I1010 13:32:24.981884 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:25 crc kubenswrapper[4745]: I1010 13:32:25.062504 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-cm4r8"] Oct 10 13:32:25 crc kubenswrapper[4745]: I1010 13:32:25.100056 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" podUID="5d9cada8-cb95-44d1-846b-0b415548ef15" containerName="dnsmasq-dns" containerID="cri-o://16eac1ccf9dbce2c4bc9030d0e822e406f75268d5ca2bbe3222daace4a4eab25" gracePeriod=10 Oct 10 13:32:25 crc kubenswrapper[4745]: I1010 13:32:25.531937 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:25 crc kubenswrapper[4745]: I1010 13:32:25.634248 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-ovsdbserver-nb\") pod \"5d9cada8-cb95-44d1-846b-0b415548ef15\" (UID: \"5d9cada8-cb95-44d1-846b-0b415548ef15\") " Oct 10 13:32:25 crc kubenswrapper[4745]: I1010 13:32:25.634312 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-config\") pod \"5d9cada8-cb95-44d1-846b-0b415548ef15\" (UID: \"5d9cada8-cb95-44d1-846b-0b415548ef15\") " Oct 10 13:32:25 crc kubenswrapper[4745]: I1010 13:32:25.634465 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-dns-svc\") pod \"5d9cada8-cb95-44d1-846b-0b415548ef15\" (UID: \"5d9cada8-cb95-44d1-846b-0b415548ef15\") " Oct 10 13:32:25 crc kubenswrapper[4745]: I1010 13:32:25.634538 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22rzr\" (UniqueName: \"kubernetes.io/projected/5d9cada8-cb95-44d1-846b-0b415548ef15-kube-api-access-22rzr\") pod \"5d9cada8-cb95-44d1-846b-0b415548ef15\" (UID: \"5d9cada8-cb95-44d1-846b-0b415548ef15\") " Oct 10 13:32:25 crc kubenswrapper[4745]: I1010 13:32:25.638870 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d9cada8-cb95-44d1-846b-0b415548ef15-kube-api-access-22rzr" (OuterVolumeSpecName: "kube-api-access-22rzr") pod "5d9cada8-cb95-44d1-846b-0b415548ef15" (UID: "5d9cada8-cb95-44d1-846b-0b415548ef15"). InnerVolumeSpecName "kube-api-access-22rzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:32:25 crc kubenswrapper[4745]: I1010 13:32:25.666834 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5d9cada8-cb95-44d1-846b-0b415548ef15" (UID: "5d9cada8-cb95-44d1-846b-0b415548ef15"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:25 crc kubenswrapper[4745]: I1010 13:32:25.668184 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5d9cada8-cb95-44d1-846b-0b415548ef15" (UID: "5d9cada8-cb95-44d1-846b-0b415548ef15"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:25 crc kubenswrapper[4745]: I1010 13:32:25.670534 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-config" (OuterVolumeSpecName: "config") pod "5d9cada8-cb95-44d1-846b-0b415548ef15" (UID: "5d9cada8-cb95-44d1-846b-0b415548ef15"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:25 crc kubenswrapper[4745]: I1010 13:32:25.736109 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22rzr\" (UniqueName: \"kubernetes.io/projected/5d9cada8-cb95-44d1-846b-0b415548ef15-kube-api-access-22rzr\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:25 crc kubenswrapper[4745]: I1010 13:32:25.736348 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:25 crc kubenswrapper[4745]: I1010 13:32:25.736420 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:25 crc kubenswrapper[4745]: I1010 13:32:25.736476 4745 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d9cada8-cb95-44d1-846b-0b415548ef15-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.107909 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-xhz46" event={"ID":"3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce","Type":"ContainerStarted","Data":"6f0ec202e32a8e08dd0e7abfd482a3bc91f91a224bd929912fc79867eb886d25"} Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.110627 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c22070ab-5d89-42fa-b04a-640ec3045982","Type":"ContainerStarted","Data":"6a0713b0ceb5dfca283aeecd3c0aa585edff074276f12ca759c210c9ce5a142d"} Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.112513 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"3894eba1-5402-4b7f-9859-a57827e29f2b","Type":"ContainerStarted","Data":"883ba645d6bb35b2a18613146ffb2e86e10c3f24b1606de6b36bcc591c45eaf4"} Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.114119 4745 generic.go:334] "Generic (PLEG): container finished" podID="5d9cada8-cb95-44d1-846b-0b415548ef15" containerID="16eac1ccf9dbce2c4bc9030d0e822e406f75268d5ca2bbe3222daace4a4eab25" exitCode=0 Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.114168 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" event={"ID":"5d9cada8-cb95-44d1-846b-0b415548ef15","Type":"ContainerDied","Data":"16eac1ccf9dbce2c4bc9030d0e822e406f75268d5ca2bbe3222daace4a4eab25"} Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.114187 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" event={"ID":"5d9cada8-cb95-44d1-846b-0b415548ef15","Type":"ContainerDied","Data":"00fc0998a0b80a514320703f3f2353d518a6df07138c534f71e68bfab1278235"} Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.114204 4745 scope.go:117] "RemoveContainer" containerID="16eac1ccf9dbce2c4bc9030d0e822e406f75268d5ca2bbe3222daace4a4eab25" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.114196 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-cm4r8" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.116767 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"dd428c33-2518-41ce-8dd4-b62fe3059525","Type":"ContainerStarted","Data":"f6c4941ef1d49bef0a9d40983f7e482417fb8465b20fbc81cdb0eee443c050a9"} Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.118526 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"35e75435-406e-4053-a1c2-2a7fc992b31e","Type":"ContainerStarted","Data":"073191204be147feb23ca595be87073751c1ee3bf4507874ec2b90a6d77573c1"} Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.118773 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-x77sv" podUID="335ddca3-4773-4797-85b2-49b959c06fc3" containerName="registry-server" containerID="cri-o://0391359040ba5c79e7934d752a64c379fb7178bdc17249455733648e28b682d1" gracePeriod=2 Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.137676 4745 scope.go:117] "RemoveContainer" containerID="a29ab365ae90f38c4f1b24ec0ca6da87542866533e565cfa9258d73766deb283" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.164569 4745 scope.go:117] "RemoveContainer" containerID="16eac1ccf9dbce2c4bc9030d0e822e406f75268d5ca2bbe3222daace4a4eab25" Oct 10 13:32:26 crc kubenswrapper[4745]: E1010 13:32:26.165965 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16eac1ccf9dbce2c4bc9030d0e822e406f75268d5ca2bbe3222daace4a4eab25\": container with ID starting with 16eac1ccf9dbce2c4bc9030d0e822e406f75268d5ca2bbe3222daace4a4eab25 not found: ID does not exist" containerID="16eac1ccf9dbce2c4bc9030d0e822e406f75268d5ca2bbe3222daace4a4eab25" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.166007 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16eac1ccf9dbce2c4bc9030d0e822e406f75268d5ca2bbe3222daace4a4eab25"} err="failed to get container status \"16eac1ccf9dbce2c4bc9030d0e822e406f75268d5ca2bbe3222daace4a4eab25\": rpc error: code = NotFound desc = could not find container \"16eac1ccf9dbce2c4bc9030d0e822e406f75268d5ca2bbe3222daace4a4eab25\": container with ID starting with 16eac1ccf9dbce2c4bc9030d0e822e406f75268d5ca2bbe3222daace4a4eab25 not found: ID does not exist" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.166033 4745 scope.go:117] "RemoveContainer" containerID="a29ab365ae90f38c4f1b24ec0ca6da87542866533e565cfa9258d73766deb283" Oct 10 13:32:26 crc kubenswrapper[4745]: E1010 13:32:26.168063 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a29ab365ae90f38c4f1b24ec0ca6da87542866533e565cfa9258d73766deb283\": container with ID starting with a29ab365ae90f38c4f1b24ec0ca6da87542866533e565cfa9258d73766deb283 not found: ID does not exist" containerID="a29ab365ae90f38c4f1b24ec0ca6da87542866533e565cfa9258d73766deb283" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.168129 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a29ab365ae90f38c4f1b24ec0ca6da87542866533e565cfa9258d73766deb283"} err="failed to get container status \"a29ab365ae90f38c4f1b24ec0ca6da87542866533e565cfa9258d73766deb283\": rpc error: code = NotFound desc = could not find container \"a29ab365ae90f38c4f1b24ec0ca6da87542866533e565cfa9258d73766deb283\": container with ID starting with a29ab365ae90f38c4f1b24ec0ca6da87542866533e565cfa9258d73766deb283 not found: ID does not exist" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.168474 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=20.869842427000002 podStartE2EDuration="29.168463239s" podCreationTimestamp="2025-10-10 13:31:57 +0000 UTC" firstStartedPulling="2025-10-10 13:32:07.401948569 +0000 UTC m=+841.299605332" lastFinishedPulling="2025-10-10 13:32:15.700569341 +0000 UTC m=+849.598226144" observedRunningTime="2025-10-10 13:32:26.164363345 +0000 UTC m=+860.062020118" watchObservedRunningTime="2025-10-10 13:32:26.168463239 +0000 UTC m=+860.066120002" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.173397 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-xhz46" podStartSLOduration=7.985383425 podStartE2EDuration="17.173388733s" podCreationTimestamp="2025-10-10 13:32:09 +0000 UTC" firstStartedPulling="2025-10-10 13:32:16.195001711 +0000 UTC m=+850.092658484" lastFinishedPulling="2025-10-10 13:32:25.383007029 +0000 UTC m=+859.280663792" observedRunningTime="2025-10-10 13:32:26.130570793 +0000 UTC m=+860.028227556" watchObservedRunningTime="2025-10-10 13:32:26.173388733 +0000 UTC m=+860.071045496" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.193212 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.473979357 podStartE2EDuration="19.193195162s" podCreationTimestamp="2025-10-10 13:32:07 +0000 UTC" firstStartedPulling="2025-10-10 13:32:10.641572584 +0000 UTC m=+844.539229347" lastFinishedPulling="2025-10-10 13:32:25.360788359 +0000 UTC m=+859.258445152" observedRunningTime="2025-10-10 13:32:26.183362564 +0000 UTC m=+860.081019367" watchObservedRunningTime="2025-10-10 13:32:26.193195162 +0000 UTC m=+860.090851925" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.210883 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=22.296664477 podStartE2EDuration="31.210864808s" podCreationTimestamp="2025-10-10 13:31:55 +0000 UTC" firstStartedPulling="2025-10-10 13:32:07.188929433 +0000 UTC m=+841.086586196" lastFinishedPulling="2025-10-10 13:32:16.103129764 +0000 UTC m=+850.000786527" observedRunningTime="2025-10-10 13:32:26.209852782 +0000 UTC m=+860.107509565" watchObservedRunningTime="2025-10-10 13:32:26.210864808 +0000 UTC m=+860.108521581" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.228546 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=5.377671458 podStartE2EDuration="23.228523083s" podCreationTimestamp="2025-10-10 13:32:03 +0000 UTC" firstStartedPulling="2025-10-10 13:32:07.550022745 +0000 UTC m=+841.447679508" lastFinishedPulling="2025-10-10 13:32:25.40087436 +0000 UTC m=+859.298531133" observedRunningTime="2025-10-10 13:32:26.227546869 +0000 UTC m=+860.125203632" watchObservedRunningTime="2025-10-10 13:32:26.228523083 +0000 UTC m=+860.126179876" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.235866 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.257794 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-cm4r8"] Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.265373 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-cm4r8"] Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.280430 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.429027 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.470968 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.540794 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x77sv" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.651640 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkzrv\" (UniqueName: \"kubernetes.io/projected/335ddca3-4773-4797-85b2-49b959c06fc3-kube-api-access-vkzrv\") pod \"335ddca3-4773-4797-85b2-49b959c06fc3\" (UID: \"335ddca3-4773-4797-85b2-49b959c06fc3\") " Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.651861 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335ddca3-4773-4797-85b2-49b959c06fc3-catalog-content\") pod \"335ddca3-4773-4797-85b2-49b959c06fc3\" (UID: \"335ddca3-4773-4797-85b2-49b959c06fc3\") " Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.651948 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335ddca3-4773-4797-85b2-49b959c06fc3-utilities\") pod \"335ddca3-4773-4797-85b2-49b959c06fc3\" (UID: \"335ddca3-4773-4797-85b2-49b959c06fc3\") " Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.652863 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/335ddca3-4773-4797-85b2-49b959c06fc3-utilities" (OuterVolumeSpecName: "utilities") pod "335ddca3-4773-4797-85b2-49b959c06fc3" (UID: "335ddca3-4773-4797-85b2-49b959c06fc3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.655681 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/335ddca3-4773-4797-85b2-49b959c06fc3-kube-api-access-vkzrv" (OuterVolumeSpecName: "kube-api-access-vkzrv") pod "335ddca3-4773-4797-85b2-49b959c06fc3" (UID: "335ddca3-4773-4797-85b2-49b959c06fc3"). InnerVolumeSpecName "kube-api-access-vkzrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.702553 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/335ddca3-4773-4797-85b2-49b959c06fc3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "335ddca3-4773-4797-85b2-49b959c06fc3" (UID: "335ddca3-4773-4797-85b2-49b959c06fc3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.765945 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkzrv\" (UniqueName: \"kubernetes.io/projected/335ddca3-4773-4797-85b2-49b959c06fc3-kube-api-access-vkzrv\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.765987 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335ddca3-4773-4797-85b2-49b959c06fc3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.766003 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335ddca3-4773-4797-85b2-49b959c06fc3-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.779709 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d9cada8-cb95-44d1-846b-0b415548ef15" path="/var/lib/kubelet/pods/5d9cada8-cb95-44d1-846b-0b415548ef15/volumes" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.909355 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 10 13:32:26 crc kubenswrapper[4745]: I1010 13:32:26.909538 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.130446 4745 generic.go:334] "Generic (PLEG): container finished" podID="335ddca3-4773-4797-85b2-49b959c06fc3" containerID="0391359040ba5c79e7934d752a64c379fb7178bdc17249455733648e28b682d1" exitCode=0 Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.130535 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x77sv" event={"ID":"335ddca3-4773-4797-85b2-49b959c06fc3","Type":"ContainerDied","Data":"0391359040ba5c79e7934d752a64c379fb7178bdc17249455733648e28b682d1"} Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.130546 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x77sv" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.130574 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x77sv" event={"ID":"335ddca3-4773-4797-85b2-49b959c06fc3","Type":"ContainerDied","Data":"ae566cd25a2985e83a37d16c011128e28fb61914353f1806469a7364a7210b59"} Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.130600 4745 scope.go:117] "RemoveContainer" containerID="0391359040ba5c79e7934d752a64c379fb7178bdc17249455733648e28b682d1" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.136584 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.136779 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.159603 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x77sv"] Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.160682 4745 scope.go:117] "RemoveContainer" containerID="b705fe2b47c8b0fc110b6996f2f3e0912d7ae653a64cfd5144261cbc5eb8e342" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.172179 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-x77sv"] Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.197552 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.203788 4745 scope.go:117] "RemoveContainer" containerID="66ca45c8e132a9c4f12b94e1888e1f77f15cdfef94962b1b59dceac5eeb319ee" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.203887 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.237361 4745 scope.go:117] "RemoveContainer" containerID="0391359040ba5c79e7934d752a64c379fb7178bdc17249455733648e28b682d1" Oct 10 13:32:27 crc kubenswrapper[4745]: E1010 13:32:27.241407 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0391359040ba5c79e7934d752a64c379fb7178bdc17249455733648e28b682d1\": container with ID starting with 0391359040ba5c79e7934d752a64c379fb7178bdc17249455733648e28b682d1 not found: ID does not exist" containerID="0391359040ba5c79e7934d752a64c379fb7178bdc17249455733648e28b682d1" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.241472 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0391359040ba5c79e7934d752a64c379fb7178bdc17249455733648e28b682d1"} err="failed to get container status \"0391359040ba5c79e7934d752a64c379fb7178bdc17249455733648e28b682d1\": rpc error: code = NotFound desc = could not find container \"0391359040ba5c79e7934d752a64c379fb7178bdc17249455733648e28b682d1\": container with ID starting with 0391359040ba5c79e7934d752a64c379fb7178bdc17249455733648e28b682d1 not found: ID does not exist" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.241616 4745 scope.go:117] "RemoveContainer" containerID="b705fe2b47c8b0fc110b6996f2f3e0912d7ae653a64cfd5144261cbc5eb8e342" Oct 10 13:32:27 crc kubenswrapper[4745]: E1010 13:32:27.252008 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b705fe2b47c8b0fc110b6996f2f3e0912d7ae653a64cfd5144261cbc5eb8e342\": container with ID starting with b705fe2b47c8b0fc110b6996f2f3e0912d7ae653a64cfd5144261cbc5eb8e342 not found: ID does not exist" containerID="b705fe2b47c8b0fc110b6996f2f3e0912d7ae653a64cfd5144261cbc5eb8e342" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.252068 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b705fe2b47c8b0fc110b6996f2f3e0912d7ae653a64cfd5144261cbc5eb8e342"} err="failed to get container status \"b705fe2b47c8b0fc110b6996f2f3e0912d7ae653a64cfd5144261cbc5eb8e342\": rpc error: code = NotFound desc = could not find container \"b705fe2b47c8b0fc110b6996f2f3e0912d7ae653a64cfd5144261cbc5eb8e342\": container with ID starting with b705fe2b47c8b0fc110b6996f2f3e0912d7ae653a64cfd5144261cbc5eb8e342 not found: ID does not exist" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.252105 4745 scope.go:117] "RemoveContainer" containerID="66ca45c8e132a9c4f12b94e1888e1f77f15cdfef94962b1b59dceac5eeb319ee" Oct 10 13:32:27 crc kubenswrapper[4745]: E1010 13:32:27.253826 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66ca45c8e132a9c4f12b94e1888e1f77f15cdfef94962b1b59dceac5eeb319ee\": container with ID starting with 66ca45c8e132a9c4f12b94e1888e1f77f15cdfef94962b1b59dceac5eeb319ee not found: ID does not exist" containerID="66ca45c8e132a9c4f12b94e1888e1f77f15cdfef94962b1b59dceac5eeb319ee" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.253855 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66ca45c8e132a9c4f12b94e1888e1f77f15cdfef94962b1b59dceac5eeb319ee"} err="failed to get container status \"66ca45c8e132a9c4f12b94e1888e1f77f15cdfef94962b1b59dceac5eeb319ee\": rpc error: code = NotFound desc = could not find container \"66ca45c8e132a9c4f12b94e1888e1f77f15cdfef94962b1b59dceac5eeb319ee\": container with ID starting with 66ca45c8e132a9c4f12b94e1888e1f77f15cdfef94962b1b59dceac5eeb319ee not found: ID does not exist" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.564174 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 10 13:32:27 crc kubenswrapper[4745]: E1010 13:32:27.564669 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea2c879a-affe-4aab-9a97-037c1257bab7" containerName="dnsmasq-dns" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.564697 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea2c879a-affe-4aab-9a97-037c1257bab7" containerName="dnsmasq-dns" Oct 10 13:32:27 crc kubenswrapper[4745]: E1010 13:32:27.564716 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d9cada8-cb95-44d1-846b-0b415548ef15" containerName="init" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.564747 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d9cada8-cb95-44d1-846b-0b415548ef15" containerName="init" Oct 10 13:32:27 crc kubenswrapper[4745]: E1010 13:32:27.564771 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="335ddca3-4773-4797-85b2-49b959c06fc3" containerName="registry-server" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.564783 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="335ddca3-4773-4797-85b2-49b959c06fc3" containerName="registry-server" Oct 10 13:32:27 crc kubenswrapper[4745]: E1010 13:32:27.564803 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="335ddca3-4773-4797-85b2-49b959c06fc3" containerName="extract-content" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.564814 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="335ddca3-4773-4797-85b2-49b959c06fc3" containerName="extract-content" Oct 10 13:32:27 crc kubenswrapper[4745]: E1010 13:32:27.564829 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d9cada8-cb95-44d1-846b-0b415548ef15" containerName="dnsmasq-dns" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.564839 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d9cada8-cb95-44d1-846b-0b415548ef15" containerName="dnsmasq-dns" Oct 10 13:32:27 crc kubenswrapper[4745]: E1010 13:32:27.564857 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="335ddca3-4773-4797-85b2-49b959c06fc3" containerName="extract-utilities" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.564868 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="335ddca3-4773-4797-85b2-49b959c06fc3" containerName="extract-utilities" Oct 10 13:32:27 crc kubenswrapper[4745]: E1010 13:32:27.564901 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69200f29-34cc-4b52-914b-25e8366dc5c7" containerName="dnsmasq-dns" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.564911 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="69200f29-34cc-4b52-914b-25e8366dc5c7" containerName="dnsmasq-dns" Oct 10 13:32:27 crc kubenswrapper[4745]: E1010 13:32:27.564928 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69200f29-34cc-4b52-914b-25e8366dc5c7" containerName="init" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.564939 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="69200f29-34cc-4b52-914b-25e8366dc5c7" containerName="init" Oct 10 13:32:27 crc kubenswrapper[4745]: E1010 13:32:27.564959 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea2c879a-affe-4aab-9a97-037c1257bab7" containerName="init" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.564968 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea2c879a-affe-4aab-9a97-037c1257bab7" containerName="init" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.565214 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="69200f29-34cc-4b52-914b-25e8366dc5c7" containerName="dnsmasq-dns" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.565242 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="335ddca3-4773-4797-85b2-49b959c06fc3" containerName="registry-server" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.565263 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d9cada8-cb95-44d1-846b-0b415548ef15" containerName="dnsmasq-dns" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.565275 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea2c879a-affe-4aab-9a97-037c1257bab7" containerName="dnsmasq-dns" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.566603 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.568271 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.568899 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-29sjm" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.569092 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.569624 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.585989 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.686245 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f12c742-f26e-473c-ab59-02217e1b24c6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.686589 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f12c742-f26e-473c-ab59-02217e1b24c6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.686633 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f12c742-f26e-473c-ab59-02217e1b24c6-config\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.686706 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f12c742-f26e-473c-ab59-02217e1b24c6-scripts\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.686743 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6f12c742-f26e-473c-ab59-02217e1b24c6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.686781 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f12c742-f26e-473c-ab59-02217e1b24c6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.686823 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtglg\" (UniqueName: \"kubernetes.io/projected/6f12c742-f26e-473c-ab59-02217e1b24c6-kube-api-access-gtglg\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.788460 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f12c742-f26e-473c-ab59-02217e1b24c6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.788506 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f12c742-f26e-473c-ab59-02217e1b24c6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.788538 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f12c742-f26e-473c-ab59-02217e1b24c6-config\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.788602 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f12c742-f26e-473c-ab59-02217e1b24c6-scripts\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.788621 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6f12c742-f26e-473c-ab59-02217e1b24c6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.788653 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f12c742-f26e-473c-ab59-02217e1b24c6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.788756 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtglg\" (UniqueName: \"kubernetes.io/projected/6f12c742-f26e-473c-ab59-02217e1b24c6-kube-api-access-gtglg\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.789152 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6f12c742-f26e-473c-ab59-02217e1b24c6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.789604 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f12c742-f26e-473c-ab59-02217e1b24c6-config\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.789644 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f12c742-f26e-473c-ab59-02217e1b24c6-scripts\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.794417 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f12c742-f26e-473c-ab59-02217e1b24c6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.794540 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f12c742-f26e-473c-ab59-02217e1b24c6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.794820 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f12c742-f26e-473c-ab59-02217e1b24c6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.817445 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtglg\" (UniqueName: \"kubernetes.io/projected/6f12c742-f26e-473c-ab59-02217e1b24c6-kube-api-access-gtglg\") pod \"ovn-northd-0\" (UID: \"6f12c742-f26e-473c-ab59-02217e1b24c6\") " pod="openstack/ovn-northd-0" Oct 10 13:32:27 crc kubenswrapper[4745]: I1010 13:32:27.892135 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 10 13:32:28 crc kubenswrapper[4745]: I1010 13:32:28.325502 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 10 13:32:28 crc kubenswrapper[4745]: I1010 13:32:28.406054 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 10 13:32:28 crc kubenswrapper[4745]: I1010 13:32:28.406092 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 10 13:32:28 crc kubenswrapper[4745]: I1010 13:32:28.759775 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="335ddca3-4773-4797-85b2-49b959c06fc3" path="/var/lib/kubelet/pods/335ddca3-4773-4797-85b2-49b959c06fc3/volumes" Oct 10 13:32:29 crc kubenswrapper[4745]: I1010 13:32:29.162691 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"6f12c742-f26e-473c-ab59-02217e1b24c6","Type":"ContainerStarted","Data":"d6bbb746090591d4a3903e00f5a68bad41d036ac019e3c2eeca90cbaefdcfa6c"} Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.176221 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"6f12c742-f26e-473c-ab59-02217e1b24c6","Type":"ContainerStarted","Data":"d8a9b81a1b4cee8bee84ec72139812ed200f42cfdca976aed0ef9de05a0449cc"} Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.176599 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"6f12c742-f26e-473c-ab59-02217e1b24c6","Type":"ContainerStarted","Data":"83d9b932e0ca3190d05670be01c6f8c43773ba3df0880a24ccb5fc5a52e24ac9"} Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.176628 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.207284 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.9852448759999999 podStartE2EDuration="3.207267206s" podCreationTimestamp="2025-10-10 13:32:27 +0000 UTC" firstStartedPulling="2025-10-10 13:32:28.340019624 +0000 UTC m=+862.237676427" lastFinishedPulling="2025-10-10 13:32:29.562041994 +0000 UTC m=+863.459698757" observedRunningTime="2025-10-10 13:32:30.206510347 +0000 UTC m=+864.104167150" watchObservedRunningTime="2025-10-10 13:32:30.207267206 +0000 UTC m=+864.104923969" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.247033 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-2vx66"] Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.249210 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.276405 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2vx66"] Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.309976 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.335146 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-2vx66\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.335236 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-dns-svc\") pod \"dnsmasq-dns-698758b865-2vx66\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.335262 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-2vx66\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.335277 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfp87\" (UniqueName: \"kubernetes.io/projected/ece46492-a015-4207-a04e-856390729ecb-kube-api-access-pfp87\") pod \"dnsmasq-dns-698758b865-2vx66\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.335326 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-config\") pod \"dnsmasq-dns-698758b865-2vx66\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.436317 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-dns-svc\") pod \"dnsmasq-dns-698758b865-2vx66\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.436645 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-2vx66\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.436662 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfp87\" (UniqueName: \"kubernetes.io/projected/ece46492-a015-4207-a04e-856390729ecb-kube-api-access-pfp87\") pod \"dnsmasq-dns-698758b865-2vx66\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.436712 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-config\") pod \"dnsmasq-dns-698758b865-2vx66\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.436758 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-2vx66\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.437443 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-dns-svc\") pod \"dnsmasq-dns-698758b865-2vx66\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.437523 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-2vx66\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.437679 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-2vx66\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.437841 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-config\") pod \"dnsmasq-dns-698758b865-2vx66\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.453971 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfp87\" (UniqueName: \"kubernetes.io/projected/ece46492-a015-4207-a04e-856390729ecb-kube-api-access-pfp87\") pod \"dnsmasq-dns-698758b865-2vx66\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.471984 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.519036 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 10 13:32:30 crc kubenswrapper[4745]: I1010 13:32:30.571079 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.039547 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2vx66"] Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.185392 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2vx66" event={"ID":"ece46492-a015-4207-a04e-856390729ecb","Type":"ContainerStarted","Data":"7c695ed56ce50e5f7407f1113f70ad6c3677d5b9d6b8f23231646d1e1fc9e6e0"} Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.365879 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.377828 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.379301 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-fs7tk" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.379542 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.380045 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.380935 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.400422 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.451489 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.451550 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.451614 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/00893894-3690-4a11-a2f7-0d1dc6c14e86-cache\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.451656 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gplcs\" (UniqueName: \"kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-kube-api-access-gplcs\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.451688 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/00893894-3690-4a11-a2f7-0d1dc6c14e86-lock\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.554012 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.554088 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.554136 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/00893894-3690-4a11-a2f7-0d1dc6c14e86-cache\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.554173 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gplcs\" (UniqueName: \"kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-kube-api-access-gplcs\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.554211 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/00893894-3690-4a11-a2f7-0d1dc6c14e86-lock\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:31 crc kubenswrapper[4745]: E1010 13:32:31.554217 4745 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 10 13:32:31 crc kubenswrapper[4745]: E1010 13:32:31.554243 4745 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 10 13:32:31 crc kubenswrapper[4745]: E1010 13:32:31.554301 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift podName:00893894-3690-4a11-a2f7-0d1dc6c14e86 nodeName:}" failed. No retries permitted until 2025-10-10 13:32:32.054282757 +0000 UTC m=+865.951939510 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift") pod "swift-storage-0" (UID: "00893894-3690-4a11-a2f7-0d1dc6c14e86") : configmap "swift-ring-files" not found Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.554538 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/swift-storage-0" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.554836 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/00893894-3690-4a11-a2f7-0d1dc6c14e86-lock\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.554890 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/00893894-3690-4a11-a2f7-0d1dc6c14e86-cache\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.583662 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gplcs\" (UniqueName: \"kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-kube-api-access-gplcs\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.588336 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.613537 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-8kzzn"] Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.614756 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.616986 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.617490 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.618249 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.625033 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-8kzzn"] Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.656179 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/749bd815-8704-47d6-b5a0-4e129ab976c1-etc-swift\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.656395 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-dispersionconf\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.656528 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/749bd815-8704-47d6-b5a0-4e129ab976c1-ring-data-devices\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.656661 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpcf2\" (UniqueName: \"kubernetes.io/projected/749bd815-8704-47d6-b5a0-4e129ab976c1-kube-api-access-vpcf2\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.656829 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/749bd815-8704-47d6-b5a0-4e129ab976c1-scripts\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.656970 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-combined-ca-bundle\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.657111 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-swiftconf\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.758338 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-dispersionconf\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.758647 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/749bd815-8704-47d6-b5a0-4e129ab976c1-ring-data-devices\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.758801 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpcf2\" (UniqueName: \"kubernetes.io/projected/749bd815-8704-47d6-b5a0-4e129ab976c1-kube-api-access-vpcf2\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.758918 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/749bd815-8704-47d6-b5a0-4e129ab976c1-scripts\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.759119 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-combined-ca-bundle\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.759242 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-swiftconf\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.759325 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/749bd815-8704-47d6-b5a0-4e129ab976c1-etc-swift\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.759899 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/749bd815-8704-47d6-b5a0-4e129ab976c1-etc-swift\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.759946 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/749bd815-8704-47d6-b5a0-4e129ab976c1-ring-data-devices\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.760465 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/749bd815-8704-47d6-b5a0-4e129ab976c1-scripts\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.762271 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-combined-ca-bundle\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.764426 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-swiftconf\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.766191 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-dispersionconf\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.788585 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpcf2\" (UniqueName: \"kubernetes.io/projected/749bd815-8704-47d6-b5a0-4e129ab976c1-kube-api-access-vpcf2\") pod \"swift-ring-rebalance-8kzzn\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:31 crc kubenswrapper[4745]: I1010 13:32:31.959666 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:32 crc kubenswrapper[4745]: I1010 13:32:32.063924 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:32 crc kubenswrapper[4745]: E1010 13:32:32.064485 4745 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 10 13:32:32 crc kubenswrapper[4745]: E1010 13:32:32.064503 4745 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 10 13:32:32 crc kubenswrapper[4745]: E1010 13:32:32.064557 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift podName:00893894-3690-4a11-a2f7-0d1dc6c14e86 nodeName:}" failed. No retries permitted until 2025-10-10 13:32:33.064540806 +0000 UTC m=+866.962197579 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift") pod "swift-storage-0" (UID: "00893894-3690-4a11-a2f7-0d1dc6c14e86") : configmap "swift-ring-files" not found Oct 10 13:32:32 crc kubenswrapper[4745]: I1010 13:32:32.202359 4745 generic.go:334] "Generic (PLEG): container finished" podID="ece46492-a015-4207-a04e-856390729ecb" containerID="4b53735422140493a15987542129134325505b81d926205f1d367d29361f88a5" exitCode=0 Oct 10 13:32:32 crc kubenswrapper[4745]: I1010 13:32:32.202403 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2vx66" event={"ID":"ece46492-a015-4207-a04e-856390729ecb","Type":"ContainerDied","Data":"4b53735422140493a15987542129134325505b81d926205f1d367d29361f88a5"} Oct 10 13:32:32 crc kubenswrapper[4745]: I1010 13:32:32.419894 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-8kzzn"] Oct 10 13:32:32 crc kubenswrapper[4745]: W1010 13:32:32.422565 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod749bd815_8704_47d6_b5a0_4e129ab976c1.slice/crio-d16f629ed3c26a6853e895071e0f82b1ecce21d1368c893d926798290ebe3e78 WatchSource:0}: Error finding container d16f629ed3c26a6853e895071e0f82b1ecce21d1368c893d926798290ebe3e78: Status 404 returned error can't find the container with id d16f629ed3c26a6853e895071e0f82b1ecce21d1368c893d926798290ebe3e78 Oct 10 13:32:33 crc kubenswrapper[4745]: I1010 13:32:33.005402 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 10 13:32:33 crc kubenswrapper[4745]: I1010 13:32:33.073468 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 10 13:32:33 crc kubenswrapper[4745]: I1010 13:32:33.084373 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:33 crc kubenswrapper[4745]: E1010 13:32:33.084544 4745 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 10 13:32:33 crc kubenswrapper[4745]: E1010 13:32:33.084576 4745 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 10 13:32:33 crc kubenswrapper[4745]: E1010 13:32:33.084639 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift podName:00893894-3690-4a11-a2f7-0d1dc6c14e86 nodeName:}" failed. No retries permitted until 2025-10-10 13:32:35.084621652 +0000 UTC m=+868.982278415 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift") pod "swift-storage-0" (UID: "00893894-3690-4a11-a2f7-0d1dc6c14e86") : configmap "swift-ring-files" not found Oct 10 13:32:33 crc kubenswrapper[4745]: I1010 13:32:33.210062 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-8kzzn" event={"ID":"749bd815-8704-47d6-b5a0-4e129ab976c1","Type":"ContainerStarted","Data":"d16f629ed3c26a6853e895071e0f82b1ecce21d1368c893d926798290ebe3e78"} Oct 10 13:32:33 crc kubenswrapper[4745]: I1010 13:32:33.212005 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2vx66" event={"ID":"ece46492-a015-4207-a04e-856390729ecb","Type":"ContainerStarted","Data":"4598efad6dd2c1f942c5a23679903d92cd3c31b4618306670f3898369a166830"} Oct 10 13:32:33 crc kubenswrapper[4745]: I1010 13:32:33.233461 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-2vx66" podStartSLOduration=3.233448126 podStartE2EDuration="3.233448126s" podCreationTimestamp="2025-10-10 13:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:32:33.228558162 +0000 UTC m=+867.126214915" watchObservedRunningTime="2025-10-10 13:32:33.233448126 +0000 UTC m=+867.131104889" Oct 10 13:32:33 crc kubenswrapper[4745]: I1010 13:32:33.897674 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-dtjmj"] Oct 10 13:32:33 crc kubenswrapper[4745]: I1010 13:32:33.902065 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dtjmj" Oct 10 13:32:33 crc kubenswrapper[4745]: I1010 13:32:33.931188 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dtjmj"] Oct 10 13:32:33 crc kubenswrapper[4745]: I1010 13:32:33.998997 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nldsj\" (UniqueName: \"kubernetes.io/projected/8c7311f9-2a50-444c-b90d-c86b0d6f887b-kube-api-access-nldsj\") pod \"glance-db-create-dtjmj\" (UID: \"8c7311f9-2a50-444c-b90d-c86b0d6f887b\") " pod="openstack/glance-db-create-dtjmj" Oct 10 13:32:34 crc kubenswrapper[4745]: I1010 13:32:34.100391 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nldsj\" (UniqueName: \"kubernetes.io/projected/8c7311f9-2a50-444c-b90d-c86b0d6f887b-kube-api-access-nldsj\") pod \"glance-db-create-dtjmj\" (UID: \"8c7311f9-2a50-444c-b90d-c86b0d6f887b\") " pod="openstack/glance-db-create-dtjmj" Oct 10 13:32:34 crc kubenswrapper[4745]: I1010 13:32:34.117054 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nldsj\" (UniqueName: \"kubernetes.io/projected/8c7311f9-2a50-444c-b90d-c86b0d6f887b-kube-api-access-nldsj\") pod \"glance-db-create-dtjmj\" (UID: \"8c7311f9-2a50-444c-b90d-c86b0d6f887b\") " pod="openstack/glance-db-create-dtjmj" Oct 10 13:32:34 crc kubenswrapper[4745]: I1010 13:32:34.220480 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:34 crc kubenswrapper[4745]: I1010 13:32:34.224613 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dtjmj" Oct 10 13:32:35 crc kubenswrapper[4745]: I1010 13:32:35.115100 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:35 crc kubenswrapper[4745]: E1010 13:32:35.115282 4745 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 10 13:32:35 crc kubenswrapper[4745]: E1010 13:32:35.115522 4745 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 10 13:32:35 crc kubenswrapper[4745]: E1010 13:32:35.115623 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift podName:00893894-3690-4a11-a2f7-0d1dc6c14e86 nodeName:}" failed. No retries permitted until 2025-10-10 13:32:39.115605963 +0000 UTC m=+873.013262726 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift") pod "swift-storage-0" (UID: "00893894-3690-4a11-a2f7-0d1dc6c14e86") : configmap "swift-ring-files" not found Oct 10 13:32:36 crc kubenswrapper[4745]: I1010 13:32:36.136013 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dtjmj"] Oct 10 13:32:36 crc kubenswrapper[4745]: W1010 13:32:36.144302 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c7311f9_2a50_444c_b90d_c86b0d6f887b.slice/crio-58bf21e7af41c37ca2d5add4820a3c2d3b4f07c4da4bddf9c9a0fcffd6b982a7 WatchSource:0}: Error finding container 58bf21e7af41c37ca2d5add4820a3c2d3b4f07c4da4bddf9c9a0fcffd6b982a7: Status 404 returned error can't find the container with id 58bf21e7af41c37ca2d5add4820a3c2d3b4f07c4da4bddf9c9a0fcffd6b982a7 Oct 10 13:32:36 crc kubenswrapper[4745]: I1010 13:32:36.236997 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-8kzzn" event={"ID":"749bd815-8704-47d6-b5a0-4e129ab976c1","Type":"ContainerStarted","Data":"f3853210d27b829d73a264611a1a8af3b60a6126ab4815a5f37f413de4ecc963"} Oct 10 13:32:36 crc kubenswrapper[4745]: I1010 13:32:36.238533 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dtjmj" event={"ID":"8c7311f9-2a50-444c-b90d-c86b0d6f887b","Type":"ContainerStarted","Data":"58bf21e7af41c37ca2d5add4820a3c2d3b4f07c4da4bddf9c9a0fcffd6b982a7"} Oct 10 13:32:36 crc kubenswrapper[4745]: I1010 13:32:36.270587 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-8kzzn" podStartSLOduration=1.956979984 podStartE2EDuration="5.270567291s" podCreationTimestamp="2025-10-10 13:32:31 +0000 UTC" firstStartedPulling="2025-10-10 13:32:32.424366381 +0000 UTC m=+866.322023134" lastFinishedPulling="2025-10-10 13:32:35.737953668 +0000 UTC m=+869.635610441" observedRunningTime="2025-10-10 13:32:36.255648404 +0000 UTC m=+870.153305177" watchObservedRunningTime="2025-10-10 13:32:36.270567291 +0000 UTC m=+870.168224054" Oct 10 13:32:37 crc kubenswrapper[4745]: I1010 13:32:37.247369 4745 generic.go:334] "Generic (PLEG): container finished" podID="8c7311f9-2a50-444c-b90d-c86b0d6f887b" containerID="c4820afa2c40ee2f1bab3f2caac896b75a46ac51298b9e561885f1d740592ddf" exitCode=0 Oct 10 13:32:37 crc kubenswrapper[4745]: I1010 13:32:37.247444 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dtjmj" event={"ID":"8c7311f9-2a50-444c-b90d-c86b0d6f887b","Type":"ContainerDied","Data":"c4820afa2c40ee2f1bab3f2caac896b75a46ac51298b9e561885f1d740592ddf"} Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.231509 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-nb8v8"] Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.233916 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nb8v8" Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.245665 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-nb8v8"] Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.370121 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sblq\" (UniqueName: \"kubernetes.io/projected/805bb2a1-bff0-4fda-be96-766d55898d4a-kube-api-access-7sblq\") pod \"keystone-db-create-nb8v8\" (UID: \"805bb2a1-bff0-4fda-be96-766d55898d4a\") " pod="openstack/keystone-db-create-nb8v8" Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.471625 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sblq\" (UniqueName: \"kubernetes.io/projected/805bb2a1-bff0-4fda-be96-766d55898d4a-kube-api-access-7sblq\") pod \"keystone-db-create-nb8v8\" (UID: \"805bb2a1-bff0-4fda-be96-766d55898d4a\") " pod="openstack/keystone-db-create-nb8v8" Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.504744 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sblq\" (UniqueName: \"kubernetes.io/projected/805bb2a1-bff0-4fda-be96-766d55898d4a-kube-api-access-7sblq\") pod \"keystone-db-create-nb8v8\" (UID: \"805bb2a1-bff0-4fda-be96-766d55898d4a\") " pod="openstack/keystone-db-create-nb8v8" Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.512195 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-fxz7c"] Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.513509 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-fxz7c" Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.520495 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-fxz7c"] Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.579285 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nb8v8" Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.666614 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dtjmj" Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.674803 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ps28\" (UniqueName: \"kubernetes.io/projected/34dd525b-9501-4685-b460-9564eed74e76-kube-api-access-2ps28\") pod \"placement-db-create-fxz7c\" (UID: \"34dd525b-9501-4685-b460-9564eed74e76\") " pod="openstack/placement-db-create-fxz7c" Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.776577 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nldsj\" (UniqueName: \"kubernetes.io/projected/8c7311f9-2a50-444c-b90d-c86b0d6f887b-kube-api-access-nldsj\") pod \"8c7311f9-2a50-444c-b90d-c86b0d6f887b\" (UID: \"8c7311f9-2a50-444c-b90d-c86b0d6f887b\") " Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.777200 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ps28\" (UniqueName: \"kubernetes.io/projected/34dd525b-9501-4685-b460-9564eed74e76-kube-api-access-2ps28\") pod \"placement-db-create-fxz7c\" (UID: \"34dd525b-9501-4685-b460-9564eed74e76\") " pod="openstack/placement-db-create-fxz7c" Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.782018 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c7311f9-2a50-444c-b90d-c86b0d6f887b-kube-api-access-nldsj" (OuterVolumeSpecName: "kube-api-access-nldsj") pod "8c7311f9-2a50-444c-b90d-c86b0d6f887b" (UID: "8c7311f9-2a50-444c-b90d-c86b0d6f887b"). InnerVolumeSpecName "kube-api-access-nldsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.804130 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ps28\" (UniqueName: \"kubernetes.io/projected/34dd525b-9501-4685-b460-9564eed74e76-kube-api-access-2ps28\") pod \"placement-db-create-fxz7c\" (UID: \"34dd525b-9501-4685-b460-9564eed74e76\") " pod="openstack/placement-db-create-fxz7c" Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.879315 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nldsj\" (UniqueName: \"kubernetes.io/projected/8c7311f9-2a50-444c-b90d-c86b0d6f887b-kube-api-access-nldsj\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:38 crc kubenswrapper[4745]: I1010 13:32:38.955251 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-fxz7c" Oct 10 13:32:39 crc kubenswrapper[4745]: I1010 13:32:39.030860 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-nb8v8"] Oct 10 13:32:39 crc kubenswrapper[4745]: W1010 13:32:39.041541 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod805bb2a1_bff0_4fda_be96_766d55898d4a.slice/crio-785c578b6bdbd6a960fc5082843d1304f25b06c6c37c9a0f22a2205a98c99d60 WatchSource:0}: Error finding container 785c578b6bdbd6a960fc5082843d1304f25b06c6c37c9a0f22a2205a98c99d60: Status 404 returned error can't find the container with id 785c578b6bdbd6a960fc5082843d1304f25b06c6c37c9a0f22a2205a98c99d60 Oct 10 13:32:39 crc kubenswrapper[4745]: I1010 13:32:39.187586 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:39 crc kubenswrapper[4745]: E1010 13:32:39.187796 4745 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 10 13:32:39 crc kubenswrapper[4745]: E1010 13:32:39.187828 4745 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 10 13:32:39 crc kubenswrapper[4745]: E1010 13:32:39.187884 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift podName:00893894-3690-4a11-a2f7-0d1dc6c14e86 nodeName:}" failed. No retries permitted until 2025-10-10 13:32:47.187866914 +0000 UTC m=+881.085523677 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift") pod "swift-storage-0" (UID: "00893894-3690-4a11-a2f7-0d1dc6c14e86") : configmap "swift-ring-files" not found Oct 10 13:32:39 crc kubenswrapper[4745]: I1010 13:32:39.275183 4745 generic.go:334] "Generic (PLEG): container finished" podID="805bb2a1-bff0-4fda-be96-766d55898d4a" containerID="3a519a98f25ce3069a56ae575dbf6156951176efb68ad1b733d2dac9e80c2e8d" exitCode=0 Oct 10 13:32:39 crc kubenswrapper[4745]: I1010 13:32:39.275238 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nb8v8" event={"ID":"805bb2a1-bff0-4fda-be96-766d55898d4a","Type":"ContainerDied","Data":"3a519a98f25ce3069a56ae575dbf6156951176efb68ad1b733d2dac9e80c2e8d"} Oct 10 13:32:39 crc kubenswrapper[4745]: I1010 13:32:39.275599 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nb8v8" event={"ID":"805bb2a1-bff0-4fda-be96-766d55898d4a","Type":"ContainerStarted","Data":"785c578b6bdbd6a960fc5082843d1304f25b06c6c37c9a0f22a2205a98c99d60"} Oct 10 13:32:39 crc kubenswrapper[4745]: I1010 13:32:39.277205 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dtjmj" event={"ID":"8c7311f9-2a50-444c-b90d-c86b0d6f887b","Type":"ContainerDied","Data":"58bf21e7af41c37ca2d5add4820a3c2d3b4f07c4da4bddf9c9a0fcffd6b982a7"} Oct 10 13:32:39 crc kubenswrapper[4745]: I1010 13:32:39.277242 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58bf21e7af41c37ca2d5add4820a3c2d3b4f07c4da4bddf9c9a0fcffd6b982a7" Oct 10 13:32:39 crc kubenswrapper[4745]: I1010 13:32:39.277253 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dtjmj" Oct 10 13:32:39 crc kubenswrapper[4745]: I1010 13:32:39.450317 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-fxz7c"] Oct 10 13:32:39 crc kubenswrapper[4745]: W1010 13:32:39.460243 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34dd525b_9501_4685_b460_9564eed74e76.slice/crio-3aabc414f2c2c252a0adbbba21192e58c432853db67ec1f7419298d5c615d97a WatchSource:0}: Error finding container 3aabc414f2c2c252a0adbbba21192e58c432853db67ec1f7419298d5c615d97a: Status 404 returned error can't find the container with id 3aabc414f2c2c252a0adbbba21192e58c432853db67ec1f7419298d5c615d97a Oct 10 13:32:40 crc kubenswrapper[4745]: I1010 13:32:40.291442 4745 generic.go:334] "Generic (PLEG): container finished" podID="34dd525b-9501-4685-b460-9564eed74e76" containerID="87dffe63db1888612bcf95d6783386baf4376547f0da3816f104721cd81fb8de" exitCode=0 Oct 10 13:32:40 crc kubenswrapper[4745]: I1010 13:32:40.291530 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-fxz7c" event={"ID":"34dd525b-9501-4685-b460-9564eed74e76","Type":"ContainerDied","Data":"87dffe63db1888612bcf95d6783386baf4376547f0da3816f104721cd81fb8de"} Oct 10 13:32:40 crc kubenswrapper[4745]: I1010 13:32:40.292003 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-fxz7c" event={"ID":"34dd525b-9501-4685-b460-9564eed74e76","Type":"ContainerStarted","Data":"3aabc414f2c2c252a0adbbba21192e58c432853db67ec1f7419298d5c615d97a"} Oct 10 13:32:40 crc kubenswrapper[4745]: I1010 13:32:40.572922 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:32:40 crc kubenswrapper[4745]: I1010 13:32:40.658394 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r2kc2"] Oct 10 13:32:40 crc kubenswrapper[4745]: I1010 13:32:40.658599 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" podUID="75ea909b-a461-4053-8a51-6b804e365c4d" containerName="dnsmasq-dns" containerID="cri-o://c8726d44b9c7eeaa21f920a1b1fff755304f278c3a170417ea69f1bd0c361b67" gracePeriod=10 Oct 10 13:32:40 crc kubenswrapper[4745]: I1010 13:32:40.806707 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nb8v8" Oct 10 13:32:40 crc kubenswrapper[4745]: I1010 13:32:40.815114 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7sblq\" (UniqueName: \"kubernetes.io/projected/805bb2a1-bff0-4fda-be96-766d55898d4a-kube-api-access-7sblq\") pod \"805bb2a1-bff0-4fda-be96-766d55898d4a\" (UID: \"805bb2a1-bff0-4fda-be96-766d55898d4a\") " Oct 10 13:32:40 crc kubenswrapper[4745]: I1010 13:32:40.832896 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/805bb2a1-bff0-4fda-be96-766d55898d4a-kube-api-access-7sblq" (OuterVolumeSpecName: "kube-api-access-7sblq") pod "805bb2a1-bff0-4fda-be96-766d55898d4a" (UID: "805bb2a1-bff0-4fda-be96-766d55898d4a"). InnerVolumeSpecName "kube-api-access-7sblq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:32:40 crc kubenswrapper[4745]: I1010 13:32:40.916748 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7sblq\" (UniqueName: \"kubernetes.io/projected/805bb2a1-bff0-4fda-be96-766d55898d4a-kube-api-access-7sblq\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.128308 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.220398 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-config\") pod \"75ea909b-a461-4053-8a51-6b804e365c4d\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.220443 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bm7x5\" (UniqueName: \"kubernetes.io/projected/75ea909b-a461-4053-8a51-6b804e365c4d-kube-api-access-bm7x5\") pod \"75ea909b-a461-4053-8a51-6b804e365c4d\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.220519 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-dns-svc\") pod \"75ea909b-a461-4053-8a51-6b804e365c4d\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.220589 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-ovsdbserver-nb\") pod \"75ea909b-a461-4053-8a51-6b804e365c4d\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.220610 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-ovsdbserver-sb\") pod \"75ea909b-a461-4053-8a51-6b804e365c4d\" (UID: \"75ea909b-a461-4053-8a51-6b804e365c4d\") " Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.225670 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75ea909b-a461-4053-8a51-6b804e365c4d-kube-api-access-bm7x5" (OuterVolumeSpecName: "kube-api-access-bm7x5") pod "75ea909b-a461-4053-8a51-6b804e365c4d" (UID: "75ea909b-a461-4053-8a51-6b804e365c4d"). InnerVolumeSpecName "kube-api-access-bm7x5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.260840 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-config" (OuterVolumeSpecName: "config") pod "75ea909b-a461-4053-8a51-6b804e365c4d" (UID: "75ea909b-a461-4053-8a51-6b804e365c4d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.269563 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "75ea909b-a461-4053-8a51-6b804e365c4d" (UID: "75ea909b-a461-4053-8a51-6b804e365c4d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.277201 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "75ea909b-a461-4053-8a51-6b804e365c4d" (UID: "75ea909b-a461-4053-8a51-6b804e365c4d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.278511 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "75ea909b-a461-4053-8a51-6b804e365c4d" (UID: "75ea909b-a461-4053-8a51-6b804e365c4d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.302541 4745 generic.go:334] "Generic (PLEG): container finished" podID="75ea909b-a461-4053-8a51-6b804e365c4d" containerID="c8726d44b9c7eeaa21f920a1b1fff755304f278c3a170417ea69f1bd0c361b67" exitCode=0 Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.302614 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.302626 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" event={"ID":"75ea909b-a461-4053-8a51-6b804e365c4d","Type":"ContainerDied","Data":"c8726d44b9c7eeaa21f920a1b1fff755304f278c3a170417ea69f1bd0c361b67"} Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.302680 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-r2kc2" event={"ID":"75ea909b-a461-4053-8a51-6b804e365c4d","Type":"ContainerDied","Data":"3105a7463afc56b15d4ea8c91a0590f26dc5eebd5c62374030318a13044f4fa6"} Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.302699 4745 scope.go:117] "RemoveContainer" containerID="c8726d44b9c7eeaa21f920a1b1fff755304f278c3a170417ea69f1bd0c361b67" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.305060 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nb8v8" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.305111 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nb8v8" event={"ID":"805bb2a1-bff0-4fda-be96-766d55898d4a","Type":"ContainerDied","Data":"785c578b6bdbd6a960fc5082843d1304f25b06c6c37c9a0f22a2205a98c99d60"} Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.305150 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="785c578b6bdbd6a960fc5082843d1304f25b06c6c37c9a0f22a2205a98c99d60" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.325805 4745 scope.go:117] "RemoveContainer" containerID="d6915a05208f1c296ad397dff3c90c1ff6198ed7451ce9319fd013a76c6dc1c8" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.323158 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.325968 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.325995 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.326014 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bm7x5\" (UniqueName: \"kubernetes.io/projected/75ea909b-a461-4053-8a51-6b804e365c4d-kube-api-access-bm7x5\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.326032 4745 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75ea909b-a461-4053-8a51-6b804e365c4d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.345780 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r2kc2"] Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.350298 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r2kc2"] Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.364599 4745 scope.go:117] "RemoveContainer" containerID="c8726d44b9c7eeaa21f920a1b1fff755304f278c3a170417ea69f1bd0c361b67" Oct 10 13:32:41 crc kubenswrapper[4745]: E1010 13:32:41.365084 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8726d44b9c7eeaa21f920a1b1fff755304f278c3a170417ea69f1bd0c361b67\": container with ID starting with c8726d44b9c7eeaa21f920a1b1fff755304f278c3a170417ea69f1bd0c361b67 not found: ID does not exist" containerID="c8726d44b9c7eeaa21f920a1b1fff755304f278c3a170417ea69f1bd0c361b67" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.365125 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8726d44b9c7eeaa21f920a1b1fff755304f278c3a170417ea69f1bd0c361b67"} err="failed to get container status \"c8726d44b9c7eeaa21f920a1b1fff755304f278c3a170417ea69f1bd0c361b67\": rpc error: code = NotFound desc = could not find container \"c8726d44b9c7eeaa21f920a1b1fff755304f278c3a170417ea69f1bd0c361b67\": container with ID starting with c8726d44b9c7eeaa21f920a1b1fff755304f278c3a170417ea69f1bd0c361b67 not found: ID does not exist" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.365150 4745 scope.go:117] "RemoveContainer" containerID="d6915a05208f1c296ad397dff3c90c1ff6198ed7451ce9319fd013a76c6dc1c8" Oct 10 13:32:41 crc kubenswrapper[4745]: E1010 13:32:41.367589 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6915a05208f1c296ad397dff3c90c1ff6198ed7451ce9319fd013a76c6dc1c8\": container with ID starting with d6915a05208f1c296ad397dff3c90c1ff6198ed7451ce9319fd013a76c6dc1c8 not found: ID does not exist" containerID="d6915a05208f1c296ad397dff3c90c1ff6198ed7451ce9319fd013a76c6dc1c8" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.367614 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6915a05208f1c296ad397dff3c90c1ff6198ed7451ce9319fd013a76c6dc1c8"} err="failed to get container status \"d6915a05208f1c296ad397dff3c90c1ff6198ed7451ce9319fd013a76c6dc1c8\": rpc error: code = NotFound desc = could not find container \"d6915a05208f1c296ad397dff3c90c1ff6198ed7451ce9319fd013a76c6dc1c8\": container with ID starting with d6915a05208f1c296ad397dff3c90c1ff6198ed7451ce9319fd013a76c6dc1c8 not found: ID does not exist" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.578771 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-fxz7c" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.631399 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ps28\" (UniqueName: \"kubernetes.io/projected/34dd525b-9501-4685-b460-9564eed74e76-kube-api-access-2ps28\") pod \"34dd525b-9501-4685-b460-9564eed74e76\" (UID: \"34dd525b-9501-4685-b460-9564eed74e76\") " Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.634699 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34dd525b-9501-4685-b460-9564eed74e76-kube-api-access-2ps28" (OuterVolumeSpecName: "kube-api-access-2ps28") pod "34dd525b-9501-4685-b460-9564eed74e76" (UID: "34dd525b-9501-4685-b460-9564eed74e76"). InnerVolumeSpecName "kube-api-access-2ps28". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:32:41 crc kubenswrapper[4745]: I1010 13:32:41.732424 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ps28\" (UniqueName: \"kubernetes.io/projected/34dd525b-9501-4685-b460-9564eed74e76-kube-api-access-2ps28\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:42 crc kubenswrapper[4745]: I1010 13:32:42.316612 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-fxz7c" event={"ID":"34dd525b-9501-4685-b460-9564eed74e76","Type":"ContainerDied","Data":"3aabc414f2c2c252a0adbbba21192e58c432853db67ec1f7419298d5c615d97a"} Oct 10 13:32:42 crc kubenswrapper[4745]: I1010 13:32:42.317082 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3aabc414f2c2c252a0adbbba21192e58c432853db67ec1f7419298d5c615d97a" Oct 10 13:32:42 crc kubenswrapper[4745]: I1010 13:32:42.316653 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-fxz7c" Oct 10 13:32:42 crc kubenswrapper[4745]: I1010 13:32:42.319309 4745 generic.go:334] "Generic (PLEG): container finished" podID="749bd815-8704-47d6-b5a0-4e129ab976c1" containerID="f3853210d27b829d73a264611a1a8af3b60a6126ab4815a5f37f413de4ecc963" exitCode=0 Oct 10 13:32:42 crc kubenswrapper[4745]: I1010 13:32:42.319375 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-8kzzn" event={"ID":"749bd815-8704-47d6-b5a0-4e129ab976c1","Type":"ContainerDied","Data":"f3853210d27b829d73a264611a1a8af3b60a6126ab4815a5f37f413de4ecc963"} Oct 10 13:32:42 crc kubenswrapper[4745]: I1010 13:32:42.760342 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75ea909b-a461-4053-8a51-6b804e365c4d" path="/var/lib/kubelet/pods/75ea909b-a461-4053-8a51-6b804e365c4d/volumes" Oct 10 13:32:42 crc kubenswrapper[4745]: I1010 13:32:42.992577 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.637671 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.668994 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/749bd815-8704-47d6-b5a0-4e129ab976c1-etc-swift\") pod \"749bd815-8704-47d6-b5a0-4e129ab976c1\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.669041 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpcf2\" (UniqueName: \"kubernetes.io/projected/749bd815-8704-47d6-b5a0-4e129ab976c1-kube-api-access-vpcf2\") pod \"749bd815-8704-47d6-b5a0-4e129ab976c1\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.669077 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/749bd815-8704-47d6-b5a0-4e129ab976c1-ring-data-devices\") pod \"749bd815-8704-47d6-b5a0-4e129ab976c1\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.669166 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-dispersionconf\") pod \"749bd815-8704-47d6-b5a0-4e129ab976c1\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.669187 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/749bd815-8704-47d6-b5a0-4e129ab976c1-scripts\") pod \"749bd815-8704-47d6-b5a0-4e129ab976c1\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.669216 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-combined-ca-bundle\") pod \"749bd815-8704-47d6-b5a0-4e129ab976c1\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.669237 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-swiftconf\") pod \"749bd815-8704-47d6-b5a0-4e129ab976c1\" (UID: \"749bd815-8704-47d6-b5a0-4e129ab976c1\") " Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.670493 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/749bd815-8704-47d6-b5a0-4e129ab976c1-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "749bd815-8704-47d6-b5a0-4e129ab976c1" (UID: "749bd815-8704-47d6-b5a0-4e129ab976c1"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.670833 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/749bd815-8704-47d6-b5a0-4e129ab976c1-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "749bd815-8704-47d6-b5a0-4e129ab976c1" (UID: "749bd815-8704-47d6-b5a0-4e129ab976c1"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.676506 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/749bd815-8704-47d6-b5a0-4e129ab976c1-kube-api-access-vpcf2" (OuterVolumeSpecName: "kube-api-access-vpcf2") pod "749bd815-8704-47d6-b5a0-4e129ab976c1" (UID: "749bd815-8704-47d6-b5a0-4e129ab976c1"). InnerVolumeSpecName "kube-api-access-vpcf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.678366 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "749bd815-8704-47d6-b5a0-4e129ab976c1" (UID: "749bd815-8704-47d6-b5a0-4e129ab976c1"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.693541 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "749bd815-8704-47d6-b5a0-4e129ab976c1" (UID: "749bd815-8704-47d6-b5a0-4e129ab976c1"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.694137 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/749bd815-8704-47d6-b5a0-4e129ab976c1-scripts" (OuterVolumeSpecName: "scripts") pod "749bd815-8704-47d6-b5a0-4e129ab976c1" (UID: "749bd815-8704-47d6-b5a0-4e129ab976c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.700120 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "749bd815-8704-47d6-b5a0-4e129ab976c1" (UID: "749bd815-8704-47d6-b5a0-4e129ab976c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.770828 4745 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.770873 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/749bd815-8704-47d6-b5a0-4e129ab976c1-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.770885 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.770897 4745 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/749bd815-8704-47d6-b5a0-4e129ab976c1-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.770912 4745 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/749bd815-8704-47d6-b5a0-4e129ab976c1-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.770923 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpcf2\" (UniqueName: \"kubernetes.io/projected/749bd815-8704-47d6-b5a0-4e129ab976c1-kube-api-access-vpcf2\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:43 crc kubenswrapper[4745]: I1010 13:32:43.770938 4745 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/749bd815-8704-47d6-b5a0-4e129ab976c1-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:44 crc kubenswrapper[4745]: I1010 13:32:44.340950 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-8kzzn" event={"ID":"749bd815-8704-47d6-b5a0-4e129ab976c1","Type":"ContainerDied","Data":"d16f629ed3c26a6853e895071e0f82b1ecce21d1368c893d926798290ebe3e78"} Oct 10 13:32:44 crc kubenswrapper[4745]: I1010 13:32:44.340993 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d16f629ed3c26a6853e895071e0f82b1ecce21d1368c893d926798290ebe3e78" Oct 10 13:32:44 crc kubenswrapper[4745]: I1010 13:32:44.341010 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-8kzzn" Oct 10 13:32:46 crc kubenswrapper[4745]: I1010 13:32:46.186819 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:32:46 crc kubenswrapper[4745]: I1010 13:32:46.188292 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:32:47 crc kubenswrapper[4745]: I1010 13:32:47.234787 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:47 crc kubenswrapper[4745]: I1010 13:32:47.249413 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/00893894-3690-4a11-a2f7-0d1dc6c14e86-etc-swift\") pod \"swift-storage-0\" (UID: \"00893894-3690-4a11-a2f7-0d1dc6c14e86\") " pod="openstack/swift-storage-0" Oct 10 13:32:47 crc kubenswrapper[4745]: I1010 13:32:47.298469 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 10 13:32:47 crc kubenswrapper[4745]: I1010 13:32:47.930151 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.375994 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-eef2-account-create-qmwvh"] Oct 10 13:32:48 crc kubenswrapper[4745]: E1010 13:32:48.377144 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75ea909b-a461-4053-8a51-6b804e365c4d" containerName="dnsmasq-dns" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.377176 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="75ea909b-a461-4053-8a51-6b804e365c4d" containerName="dnsmasq-dns" Oct 10 13:32:48 crc kubenswrapper[4745]: E1010 13:32:48.377206 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75ea909b-a461-4053-8a51-6b804e365c4d" containerName="init" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.377218 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="75ea909b-a461-4053-8a51-6b804e365c4d" containerName="init" Oct 10 13:32:48 crc kubenswrapper[4745]: E1010 13:32:48.377236 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34dd525b-9501-4685-b460-9564eed74e76" containerName="mariadb-database-create" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.377249 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="34dd525b-9501-4685-b460-9564eed74e76" containerName="mariadb-database-create" Oct 10 13:32:48 crc kubenswrapper[4745]: E1010 13:32:48.377288 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c7311f9-2a50-444c-b90d-c86b0d6f887b" containerName="mariadb-database-create" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.377301 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c7311f9-2a50-444c-b90d-c86b0d6f887b" containerName="mariadb-database-create" Oct 10 13:32:48 crc kubenswrapper[4745]: E1010 13:32:48.377325 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="805bb2a1-bff0-4fda-be96-766d55898d4a" containerName="mariadb-database-create" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.377339 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="805bb2a1-bff0-4fda-be96-766d55898d4a" containerName="mariadb-database-create" Oct 10 13:32:48 crc kubenswrapper[4745]: E1010 13:32:48.377359 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="749bd815-8704-47d6-b5a0-4e129ab976c1" containerName="swift-ring-rebalance" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.377371 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="749bd815-8704-47d6-b5a0-4e129ab976c1" containerName="swift-ring-rebalance" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.377697 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="34dd525b-9501-4685-b460-9564eed74e76" containerName="mariadb-database-create" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.377716 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="749bd815-8704-47d6-b5a0-4e129ab976c1" containerName="swift-ring-rebalance" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.377772 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c7311f9-2a50-444c-b90d-c86b0d6f887b" containerName="mariadb-database-create" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.377791 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="805bb2a1-bff0-4fda-be96-766d55898d4a" containerName="mariadb-database-create" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.377814 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="75ea909b-a461-4053-8a51-6b804e365c4d" containerName="dnsmasq-dns" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.378667 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-eef2-account-create-qmwvh" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.382994 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"00893894-3690-4a11-a2f7-0d1dc6c14e86","Type":"ContainerStarted","Data":"8daed21cb335b52deeff0948a4aaae1ab01a07cb2a109e0fa0707ff4dcbeda0c"} Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.393114 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-eef2-account-create-qmwvh"] Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.403891 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.465297 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnprn\" (UniqueName: \"kubernetes.io/projected/7e42a569-8a3d-410c-8fd8-98eeee59ebe9-kube-api-access-mnprn\") pod \"keystone-eef2-account-create-qmwvh\" (UID: \"7e42a569-8a3d-410c-8fd8-98eeee59ebe9\") " pod="openstack/keystone-eef2-account-create-qmwvh" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.567630 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnprn\" (UniqueName: \"kubernetes.io/projected/7e42a569-8a3d-410c-8fd8-98eeee59ebe9-kube-api-access-mnprn\") pod \"keystone-eef2-account-create-qmwvh\" (UID: \"7e42a569-8a3d-410c-8fd8-98eeee59ebe9\") " pod="openstack/keystone-eef2-account-create-qmwvh" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.567672 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-c34c-account-create-2hp4m"] Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.569147 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c34c-account-create-2hp4m" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.571153 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.578692 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c34c-account-create-2hp4m"] Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.600017 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnprn\" (UniqueName: \"kubernetes.io/projected/7e42a569-8a3d-410c-8fd8-98eeee59ebe9-kube-api-access-mnprn\") pod \"keystone-eef2-account-create-qmwvh\" (UID: \"7e42a569-8a3d-410c-8fd8-98eeee59ebe9\") " pod="openstack/keystone-eef2-account-create-qmwvh" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.669445 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n9hl\" (UniqueName: \"kubernetes.io/projected/f136f642-8772-41b9-88d8-b28bb1111fef-kube-api-access-9n9hl\") pod \"placement-c34c-account-create-2hp4m\" (UID: \"f136f642-8772-41b9-88d8-b28bb1111fef\") " pod="openstack/placement-c34c-account-create-2hp4m" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.748610 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-eef2-account-create-qmwvh" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.771935 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n9hl\" (UniqueName: \"kubernetes.io/projected/f136f642-8772-41b9-88d8-b28bb1111fef-kube-api-access-9n9hl\") pod \"placement-c34c-account-create-2hp4m\" (UID: \"f136f642-8772-41b9-88d8-b28bb1111fef\") " pod="openstack/placement-c34c-account-create-2hp4m" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.799296 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n9hl\" (UniqueName: \"kubernetes.io/projected/f136f642-8772-41b9-88d8-b28bb1111fef-kube-api-access-9n9hl\") pod \"placement-c34c-account-create-2hp4m\" (UID: \"f136f642-8772-41b9-88d8-b28bb1111fef\") " pod="openstack/placement-c34c-account-create-2hp4m" Oct 10 13:32:48 crc kubenswrapper[4745]: I1010 13:32:48.886007 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c34c-account-create-2hp4m" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.248881 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-eef2-account-create-qmwvh"] Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.340586 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c34c-account-create-2hp4m"] Oct 10 13:32:49 crc kubenswrapper[4745]: W1010 13:32:49.385301 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e42a569_8a3d_410c_8fd8_98eeee59ebe9.slice/crio-b812e07da8e86bf3b094d7ad19648750380674f42bd998ecaf0adc0c8377628a WatchSource:0}: Error finding container b812e07da8e86bf3b094d7ad19648750380674f42bd998ecaf0adc0c8377628a: Status 404 returned error can't find the container with id b812e07da8e86bf3b094d7ad19648750380674f42bd998ecaf0adc0c8377628a Oct 10 13:32:49 crc kubenswrapper[4745]: W1010 13:32:49.388317 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf136f642_8772_41b9_88d8_b28bb1111fef.slice/crio-4eac4fc149c7df9411d48d51f42642010801749e2819e07c0ab9d10301cdf3c1 WatchSource:0}: Error finding container 4eac4fc149c7df9411d48d51f42642010801749e2819e07c0ab9d10301cdf3c1: Status 404 returned error can't find the container with id 4eac4fc149c7df9411d48d51f42642010801749e2819e07c0ab9d10301cdf3c1 Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.445446 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-92pgn" podUID="513dcce5-f68e-42a4-ba94-44013577e96e" containerName="ovn-controller" probeResult="failure" output=< Oct 10 13:32:49 crc kubenswrapper[4745]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 10 13:32:49 crc kubenswrapper[4745]: > Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.514716 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.541800 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-s2qdr" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.764759 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-92pgn-config-jnhgp"] Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.765741 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.770455 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.791067 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6fns\" (UniqueName: \"kubernetes.io/projected/1a845349-9de2-48bf-839d-cc06ec4a7619-kube-api-access-n6fns\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.791157 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a845349-9de2-48bf-839d-cc06ec4a7619-scripts\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.791220 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-run-ovn\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.791258 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-log-ovn\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.791315 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-run\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.791395 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1a845349-9de2-48bf-839d-cc06ec4a7619-additional-scripts\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.794982 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-92pgn-config-jnhgp"] Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.892433 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6fns\" (UniqueName: \"kubernetes.io/projected/1a845349-9de2-48bf-839d-cc06ec4a7619-kube-api-access-n6fns\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.892816 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a845349-9de2-48bf-839d-cc06ec4a7619-scripts\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.894603 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a845349-9de2-48bf-839d-cc06ec4a7619-scripts\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.894667 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-run-ovn\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.894700 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-log-ovn\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.894948 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-run\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.894872 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-run-ovn\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.895067 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-run\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.895125 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1a845349-9de2-48bf-839d-cc06ec4a7619-additional-scripts\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.895127 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-log-ovn\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.895523 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1a845349-9de2-48bf-839d-cc06ec4a7619-additional-scripts\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:49 crc kubenswrapper[4745]: I1010 13:32:49.910144 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6fns\" (UniqueName: \"kubernetes.io/projected/1a845349-9de2-48bf-839d-cc06ec4a7619-kube-api-access-n6fns\") pod \"ovn-controller-92pgn-config-jnhgp\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:50 crc kubenswrapper[4745]: I1010 13:32:50.165020 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:50 crc kubenswrapper[4745]: I1010 13:32:50.401334 4745 generic.go:334] "Generic (PLEG): container finished" podID="234155a2-9751-459f-b259-6bb32eb4b4d7" containerID="ba8c59cb21a1b92f873d2755789e9b4ca0999c1be1d4d2bdf990e767e13ffec7" exitCode=0 Oct 10 13:32:50 crc kubenswrapper[4745]: I1010 13:32:50.401431 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"234155a2-9751-459f-b259-6bb32eb4b4d7","Type":"ContainerDied","Data":"ba8c59cb21a1b92f873d2755789e9b4ca0999c1be1d4d2bdf990e767e13ffec7"} Oct 10 13:32:50 crc kubenswrapper[4745]: I1010 13:32:50.404389 4745 generic.go:334] "Generic (PLEG): container finished" podID="7e42a569-8a3d-410c-8fd8-98eeee59ebe9" containerID="c73d4a837753e4d6f37bf1af6f7a91fcfb0909353cef4a00bbde1222cb38bf22" exitCode=0 Oct 10 13:32:50 crc kubenswrapper[4745]: I1010 13:32:50.404465 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-eef2-account-create-qmwvh" event={"ID":"7e42a569-8a3d-410c-8fd8-98eeee59ebe9","Type":"ContainerDied","Data":"c73d4a837753e4d6f37bf1af6f7a91fcfb0909353cef4a00bbde1222cb38bf22"} Oct 10 13:32:50 crc kubenswrapper[4745]: I1010 13:32:50.404497 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-eef2-account-create-qmwvh" event={"ID":"7e42a569-8a3d-410c-8fd8-98eeee59ebe9","Type":"ContainerStarted","Data":"b812e07da8e86bf3b094d7ad19648750380674f42bd998ecaf0adc0c8377628a"} Oct 10 13:32:50 crc kubenswrapper[4745]: I1010 13:32:50.407574 4745 generic.go:334] "Generic (PLEG): container finished" podID="f136f642-8772-41b9-88d8-b28bb1111fef" containerID="0ddc1787eb1e230a614cf715c9d31bc4b65fdee8e522ab0d1fb86425f72ad96d" exitCode=0 Oct 10 13:32:50 crc kubenswrapper[4745]: I1010 13:32:50.407614 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c34c-account-create-2hp4m" event={"ID":"f136f642-8772-41b9-88d8-b28bb1111fef","Type":"ContainerDied","Data":"0ddc1787eb1e230a614cf715c9d31bc4b65fdee8e522ab0d1fb86425f72ad96d"} Oct 10 13:32:50 crc kubenswrapper[4745]: I1010 13:32:50.407650 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c34c-account-create-2hp4m" event={"ID":"f136f642-8772-41b9-88d8-b28bb1111fef","Type":"ContainerStarted","Data":"4eac4fc149c7df9411d48d51f42642010801749e2819e07c0ab9d10301cdf3c1"} Oct 10 13:32:50 crc kubenswrapper[4745]: I1010 13:32:50.411016 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"00893894-3690-4a11-a2f7-0d1dc6c14e86","Type":"ContainerStarted","Data":"2b5b3733624f2a13bfe2631b045fbf2813e9776af957be90162ffd8948e91f30"} Oct 10 13:32:50 crc kubenswrapper[4745]: I1010 13:32:50.411080 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"00893894-3690-4a11-a2f7-0d1dc6c14e86","Type":"ContainerStarted","Data":"27181b950640326f8edb92ac47c5db873c84ab16927e7b9cef3160267b93fc27"} Oct 10 13:32:50 crc kubenswrapper[4745]: I1010 13:32:50.411093 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"00893894-3690-4a11-a2f7-0d1dc6c14e86","Type":"ContainerStarted","Data":"9f20355bea2999b3065eff124fa31dada0e56a381c14de70f0d216089c00216f"} Oct 10 13:32:50 crc kubenswrapper[4745]: I1010 13:32:50.411102 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"00893894-3690-4a11-a2f7-0d1dc6c14e86","Type":"ContainerStarted","Data":"4bb1777ab1fd5381ce7e8ae7295f2db3ffd00998cd5a6413ca788122200ad14d"} Oct 10 13:32:50 crc kubenswrapper[4745]: I1010 13:32:50.666947 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-92pgn-config-jnhgp"] Oct 10 13:32:50 crc kubenswrapper[4745]: W1010 13:32:50.673394 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a845349_9de2_48bf_839d_cc06ec4a7619.slice/crio-025e3248277fe5899961ff12743ed2773f6b69e6653e492467c7cf2dbcafe739 WatchSource:0}: Error finding container 025e3248277fe5899961ff12743ed2773f6b69e6653e492467c7cf2dbcafe739: Status 404 returned error can't find the container with id 025e3248277fe5899961ff12743ed2773f6b69e6653e492467c7cf2dbcafe739 Oct 10 13:32:51 crc kubenswrapper[4745]: I1010 13:32:51.424930 4745 generic.go:334] "Generic (PLEG): container finished" podID="bc0bb94e-36a7-413d-9ce0-eea2031d1d19" containerID="e9b2c0742bac99a38637cce65f207a0590e5ea6ba54dc0c7bcb8a39f75955347" exitCode=0 Oct 10 13:32:51 crc kubenswrapper[4745]: I1010 13:32:51.424974 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bc0bb94e-36a7-413d-9ce0-eea2031d1d19","Type":"ContainerDied","Data":"e9b2c0742bac99a38637cce65f207a0590e5ea6ba54dc0c7bcb8a39f75955347"} Oct 10 13:32:51 crc kubenswrapper[4745]: I1010 13:32:51.430332 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"234155a2-9751-459f-b259-6bb32eb4b4d7","Type":"ContainerStarted","Data":"c7ea1a57fde8811da41997a7bcd0d2d8f76592e019f75dd145aad49f9c8962d6"} Oct 10 13:32:51 crc kubenswrapper[4745]: I1010 13:32:51.430547 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:32:51 crc kubenswrapper[4745]: I1010 13:32:51.434908 4745 generic.go:334] "Generic (PLEG): container finished" podID="1a845349-9de2-48bf-839d-cc06ec4a7619" containerID="65ffcca13477e662350045a787cb182472c801c2ae1b42427a3267840f703e16" exitCode=0 Oct 10 13:32:51 crc kubenswrapper[4745]: I1010 13:32:51.435057 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-92pgn-config-jnhgp" event={"ID":"1a845349-9de2-48bf-839d-cc06ec4a7619","Type":"ContainerDied","Data":"65ffcca13477e662350045a787cb182472c801c2ae1b42427a3267840f703e16"} Oct 10 13:32:51 crc kubenswrapper[4745]: I1010 13:32:51.435159 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-92pgn-config-jnhgp" event={"ID":"1a845349-9de2-48bf-839d-cc06ec4a7619","Type":"ContainerStarted","Data":"025e3248277fe5899961ff12743ed2773f6b69e6653e492467c7cf2dbcafe739"} Oct 10 13:32:51 crc kubenswrapper[4745]: I1010 13:32:51.485395 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=50.936380995 podStartE2EDuration="58.485377753s" podCreationTimestamp="2025-10-10 13:31:53 +0000 UTC" firstStartedPulling="2025-10-10 13:32:07.542350769 +0000 UTC m=+841.440007532" lastFinishedPulling="2025-10-10 13:32:15.091347527 +0000 UTC m=+848.989004290" observedRunningTime="2025-10-10 13:32:51.476327065 +0000 UTC m=+885.373983828" watchObservedRunningTime="2025-10-10 13:32:51.485377753 +0000 UTC m=+885.383034536" Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.038710 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c34c-account-create-2hp4m" Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.067451 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-eef2-account-create-qmwvh" Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.159305 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnprn\" (UniqueName: \"kubernetes.io/projected/7e42a569-8a3d-410c-8fd8-98eeee59ebe9-kube-api-access-mnprn\") pod \"7e42a569-8a3d-410c-8fd8-98eeee59ebe9\" (UID: \"7e42a569-8a3d-410c-8fd8-98eeee59ebe9\") " Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.159445 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n9hl\" (UniqueName: \"kubernetes.io/projected/f136f642-8772-41b9-88d8-b28bb1111fef-kube-api-access-9n9hl\") pod \"f136f642-8772-41b9-88d8-b28bb1111fef\" (UID: \"f136f642-8772-41b9-88d8-b28bb1111fef\") " Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.164921 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e42a569-8a3d-410c-8fd8-98eeee59ebe9-kube-api-access-mnprn" (OuterVolumeSpecName: "kube-api-access-mnprn") pod "7e42a569-8a3d-410c-8fd8-98eeee59ebe9" (UID: "7e42a569-8a3d-410c-8fd8-98eeee59ebe9"). InnerVolumeSpecName "kube-api-access-mnprn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.165866 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f136f642-8772-41b9-88d8-b28bb1111fef-kube-api-access-9n9hl" (OuterVolumeSpecName: "kube-api-access-9n9hl") pod "f136f642-8772-41b9-88d8-b28bb1111fef" (UID: "f136f642-8772-41b9-88d8-b28bb1111fef"). InnerVolumeSpecName "kube-api-access-9n9hl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.261698 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n9hl\" (UniqueName: \"kubernetes.io/projected/f136f642-8772-41b9-88d8-b28bb1111fef-kube-api-access-9n9hl\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.261753 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnprn\" (UniqueName: \"kubernetes.io/projected/7e42a569-8a3d-410c-8fd8-98eeee59ebe9-kube-api-access-mnprn\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.452565 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-eef2-account-create-qmwvh" event={"ID":"7e42a569-8a3d-410c-8fd8-98eeee59ebe9","Type":"ContainerDied","Data":"b812e07da8e86bf3b094d7ad19648750380674f42bd998ecaf0adc0c8377628a"} Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.452605 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b812e07da8e86bf3b094d7ad19648750380674f42bd998ecaf0adc0c8377628a" Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.452672 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-eef2-account-create-qmwvh" Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.457530 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bc0bb94e-36a7-413d-9ce0-eea2031d1d19","Type":"ContainerStarted","Data":"83f39747e46356d5c08ec33d69e30ab2569a1f7461903ed89ce9864a0017be49"} Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.457882 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.458854 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c34c-account-create-2hp4m" event={"ID":"f136f642-8772-41b9-88d8-b28bb1111fef","Type":"ContainerDied","Data":"4eac4fc149c7df9411d48d51f42642010801749e2819e07c0ab9d10301cdf3c1"} Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.458884 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4eac4fc149c7df9411d48d51f42642010801749e2819e07c0ab9d10301cdf3c1" Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.458944 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c34c-account-create-2hp4m" Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.478024 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"00893894-3690-4a11-a2f7-0d1dc6c14e86","Type":"ContainerStarted","Data":"4956e73cb3e9dafd26d5cbb07d8eb6f230f0a48a4ea26e3e06644c06a5c2265e"} Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.478071 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"00893894-3690-4a11-a2f7-0d1dc6c14e86","Type":"ContainerStarted","Data":"adc35f9900e08803a2b94ec5e7e6bb9a971a806051fb32a7da35ce5d8531a294"} Oct 10 13:32:52 crc kubenswrapper[4745]: I1010 13:32:52.505967 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=50.523985697 podStartE2EDuration="59.505937961s" podCreationTimestamp="2025-10-10 13:31:53 +0000 UTC" firstStartedPulling="2025-10-10 13:32:06.718614587 +0000 UTC m=+840.616271350" lastFinishedPulling="2025-10-10 13:32:15.700566851 +0000 UTC m=+849.598223614" observedRunningTime="2025-10-10 13:32:52.490602804 +0000 UTC m=+886.388259567" watchObservedRunningTime="2025-10-10 13:32:52.505937961 +0000 UTC m=+886.403594744" Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.493944 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"00893894-3690-4a11-a2f7-0d1dc6c14e86","Type":"ContainerStarted","Data":"c78fb315488d057402580ea78def9e7f27e298b87c577f72b552f71790d95a71"} Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.493989 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"00893894-3690-4a11-a2f7-0d1dc6c14e86","Type":"ContainerStarted","Data":"d1491cb99f2d362acbbee9ec2b49bd0dca553837d66264410ec8b79c648bc1d8"} Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.859067 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.899471 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-218e-account-create-l8rbx"] Oct 10 13:32:53 crc kubenswrapper[4745]: E1010 13:32:53.899853 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f136f642-8772-41b9-88d8-b28bb1111fef" containerName="mariadb-account-create" Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.899872 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="f136f642-8772-41b9-88d8-b28bb1111fef" containerName="mariadb-account-create" Oct 10 13:32:53 crc kubenswrapper[4745]: E1010 13:32:53.899891 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a845349-9de2-48bf-839d-cc06ec4a7619" containerName="ovn-config" Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.899899 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a845349-9de2-48bf-839d-cc06ec4a7619" containerName="ovn-config" Oct 10 13:32:53 crc kubenswrapper[4745]: E1010 13:32:53.899917 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e42a569-8a3d-410c-8fd8-98eeee59ebe9" containerName="mariadb-account-create" Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.899925 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e42a569-8a3d-410c-8fd8-98eeee59ebe9" containerName="mariadb-account-create" Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.900122 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="f136f642-8772-41b9-88d8-b28bb1111fef" containerName="mariadb-account-create" Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.900174 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a845349-9de2-48bf-839d-cc06ec4a7619" containerName="ovn-config" Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.900197 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e42a569-8a3d-410c-8fd8-98eeee59ebe9" containerName="mariadb-account-create" Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.900706 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-218e-account-create-l8rbx" Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.913162 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.919738 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-218e-account-create-l8rbx"] Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.998640 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-log-ovn\") pod \"1a845349-9de2-48bf-839d-cc06ec4a7619\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.998716 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1a845349-9de2-48bf-839d-cc06ec4a7619-additional-scripts\") pod \"1a845349-9de2-48bf-839d-cc06ec4a7619\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.998839 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a845349-9de2-48bf-839d-cc06ec4a7619-scripts\") pod \"1a845349-9de2-48bf-839d-cc06ec4a7619\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.998874 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-run\") pod \"1a845349-9de2-48bf-839d-cc06ec4a7619\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.998920 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-run-ovn\") pod \"1a845349-9de2-48bf-839d-cc06ec4a7619\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.998990 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6fns\" (UniqueName: \"kubernetes.io/projected/1a845349-9de2-48bf-839d-cc06ec4a7619-kube-api-access-n6fns\") pod \"1a845349-9de2-48bf-839d-cc06ec4a7619\" (UID: \"1a845349-9de2-48bf-839d-cc06ec4a7619\") " Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.999308 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m98x5\" (UniqueName: \"kubernetes.io/projected/981cf93b-d455-4364-8d16-26b5d8e8ac56-kube-api-access-m98x5\") pod \"glance-218e-account-create-l8rbx\" (UID: \"981cf93b-d455-4364-8d16-26b5d8e8ac56\") " pod="openstack/glance-218e-account-create-l8rbx" Oct 10 13:32:53 crc kubenswrapper[4745]: I1010 13:32:53.999449 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "1a845349-9de2-48bf-839d-cc06ec4a7619" (UID: "1a845349-9de2-48bf-839d-cc06ec4a7619"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.000325 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a845349-9de2-48bf-839d-cc06ec4a7619-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "1a845349-9de2-48bf-839d-cc06ec4a7619" (UID: "1a845349-9de2-48bf-839d-cc06ec4a7619"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.000347 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "1a845349-9de2-48bf-839d-cc06ec4a7619" (UID: "1a845349-9de2-48bf-839d-cc06ec4a7619"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.000318 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-run" (OuterVolumeSpecName: "var-run") pod "1a845349-9de2-48bf-839d-cc06ec4a7619" (UID: "1a845349-9de2-48bf-839d-cc06ec4a7619"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.000620 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a845349-9de2-48bf-839d-cc06ec4a7619-scripts" (OuterVolumeSpecName: "scripts") pod "1a845349-9de2-48bf-839d-cc06ec4a7619" (UID: "1a845349-9de2-48bf-839d-cc06ec4a7619"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.008694 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a845349-9de2-48bf-839d-cc06ec4a7619-kube-api-access-n6fns" (OuterVolumeSpecName: "kube-api-access-n6fns") pod "1a845349-9de2-48bf-839d-cc06ec4a7619" (UID: "1a845349-9de2-48bf-839d-cc06ec4a7619"). InnerVolumeSpecName "kube-api-access-n6fns". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.100270 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m98x5\" (UniqueName: \"kubernetes.io/projected/981cf93b-d455-4364-8d16-26b5d8e8ac56-kube-api-access-m98x5\") pod \"glance-218e-account-create-l8rbx\" (UID: \"981cf93b-d455-4364-8d16-26b5d8e8ac56\") " pod="openstack/glance-218e-account-create-l8rbx" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.100398 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6fns\" (UniqueName: \"kubernetes.io/projected/1a845349-9de2-48bf-839d-cc06ec4a7619-kube-api-access-n6fns\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.100411 4745 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.100419 4745 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1a845349-9de2-48bf-839d-cc06ec4a7619-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.100427 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a845349-9de2-48bf-839d-cc06ec4a7619-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.100437 4745 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-run\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.100445 4745 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1a845349-9de2-48bf-839d-cc06ec4a7619-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.118757 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m98x5\" (UniqueName: \"kubernetes.io/projected/981cf93b-d455-4364-8d16-26b5d8e8ac56-kube-api-access-m98x5\") pod \"glance-218e-account-create-l8rbx\" (UID: \"981cf93b-d455-4364-8d16-26b5d8e8ac56\") " pod="openstack/glance-218e-account-create-l8rbx" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.236867 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-218e-account-create-l8rbx" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.488863 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-92pgn" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.519655 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-92pgn-config-jnhgp" event={"ID":"1a845349-9de2-48bf-839d-cc06ec4a7619","Type":"ContainerDied","Data":"025e3248277fe5899961ff12743ed2773f6b69e6653e492467c7cf2dbcafe739"} Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.519691 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="025e3248277fe5899961ff12743ed2773f6b69e6653e492467c7cf2dbcafe739" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.519764 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-92pgn-config-jnhgp" Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.544653 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"00893894-3690-4a11-a2f7-0d1dc6c14e86","Type":"ContainerStarted","Data":"4605235ee767a51f12d4ac05f7c7a1c3118b844fc74bfc5d0ee5649cf02a8495"} Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.544695 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"00893894-3690-4a11-a2f7-0d1dc6c14e86","Type":"ContainerStarted","Data":"5666a71ba08fcb5db8059087ee563a4127f5bb906f953699abbea4e79ff98315"} Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.544706 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"00893894-3690-4a11-a2f7-0d1dc6c14e86","Type":"ContainerStarted","Data":"de2ded29b14ce4d3a73e8d31a469d764f0e763e2e21698e3b4aa92a3d61cc792"} Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.716373 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-218e-account-create-l8rbx"] Oct 10 13:32:54 crc kubenswrapper[4745]: W1010 13:32:54.730548 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod981cf93b_d455_4364_8d16_26b5d8e8ac56.slice/crio-a6fb852bce7a8373ac30ac03ecd70cc89c65803b7e5ba8c0a6aa6ae1273fe84f WatchSource:0}: Error finding container a6fb852bce7a8373ac30ac03ecd70cc89c65803b7e5ba8c0a6aa6ae1273fe84f: Status 404 returned error can't find the container with id a6fb852bce7a8373ac30ac03ecd70cc89c65803b7e5ba8c0a6aa6ae1273fe84f Oct 10 13:32:54 crc kubenswrapper[4745]: I1010 13:32:54.991752 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-92pgn-config-jnhgp"] Oct 10 13:32:55 crc kubenswrapper[4745]: I1010 13:32:55.003679 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-92pgn-config-jnhgp"] Oct 10 13:32:55 crc kubenswrapper[4745]: I1010 13:32:55.558965 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"00893894-3690-4a11-a2f7-0d1dc6c14e86","Type":"ContainerStarted","Data":"88ac9ecd7731e7d8bf2dac124afc69071cf8e8fa0e08797fd5b7038153e7ab00"} Oct 10 13:32:55 crc kubenswrapper[4745]: I1010 13:32:55.559011 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"00893894-3690-4a11-a2f7-0d1dc6c14e86","Type":"ContainerStarted","Data":"54ac2026e9705e4ab86ef9116bb047c031a54a0f872fddd823027dd192620b5e"} Oct 10 13:32:55 crc kubenswrapper[4745]: I1010 13:32:55.559026 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"00893894-3690-4a11-a2f7-0d1dc6c14e86","Type":"ContainerStarted","Data":"10d77cf6e0a35f6c43a67bf9c00ce7e8e75c97f52672051dc77849a3ebe80e2e"} Oct 10 13:32:55 crc kubenswrapper[4745]: I1010 13:32:55.559038 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"00893894-3690-4a11-a2f7-0d1dc6c14e86","Type":"ContainerStarted","Data":"6b9406ac560b8c4c0f3eac54ae92a7230fd983e3c16c0dd71dacb42cb9ee1868"} Oct 10 13:32:55 crc kubenswrapper[4745]: I1010 13:32:55.561973 4745 generic.go:334] "Generic (PLEG): container finished" podID="981cf93b-d455-4364-8d16-26b5d8e8ac56" containerID="2ad421cb34f2c4e036e33aa1660436c3b3bc8a66e4c156aa432e25fb99ab3257" exitCode=0 Oct 10 13:32:55 crc kubenswrapper[4745]: I1010 13:32:55.562017 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-218e-account-create-l8rbx" event={"ID":"981cf93b-d455-4364-8d16-26b5d8e8ac56","Type":"ContainerDied","Data":"2ad421cb34f2c4e036e33aa1660436c3b3bc8a66e4c156aa432e25fb99ab3257"} Oct 10 13:32:55 crc kubenswrapper[4745]: I1010 13:32:55.562047 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-218e-account-create-l8rbx" event={"ID":"981cf93b-d455-4364-8d16-26b5d8e8ac56","Type":"ContainerStarted","Data":"a6fb852bce7a8373ac30ac03ecd70cc89c65803b7e5ba8c0a6aa6ae1273fe84f"} Oct 10 13:32:55 crc kubenswrapper[4745]: I1010 13:32:55.603574 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=19.723512071000002 podStartE2EDuration="25.603558363s" podCreationTimestamp="2025-10-10 13:32:30 +0000 UTC" firstStartedPulling="2025-10-10 13:32:47.951564562 +0000 UTC m=+881.849221325" lastFinishedPulling="2025-10-10 13:32:53.831610854 +0000 UTC m=+887.729267617" observedRunningTime="2025-10-10 13:32:55.597976712 +0000 UTC m=+889.495633475" watchObservedRunningTime="2025-10-10 13:32:55.603558363 +0000 UTC m=+889.501215126" Oct 10 13:32:55 crc kubenswrapper[4745]: I1010 13:32:55.877580 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-nv8xl"] Oct 10 13:32:55 crc kubenswrapper[4745]: I1010 13:32:55.879666 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:55 crc kubenswrapper[4745]: I1010 13:32:55.882100 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 10 13:32:55 crc kubenswrapper[4745]: I1010 13:32:55.891269 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-nv8xl"] Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.034192 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbmh5\" (UniqueName: \"kubernetes.io/projected/b683a409-b7d3-404f-ac5a-43be34e071c6-kube-api-access-mbmh5\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.034549 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.034591 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-config\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.034608 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.034628 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.034668 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.136497 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.136584 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.136650 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbmh5\" (UniqueName: \"kubernetes.io/projected/b683a409-b7d3-404f-ac5a-43be34e071c6-kube-api-access-mbmh5\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.136698 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.136761 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-config\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.136777 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.137678 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.138274 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.138899 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.140289 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-config\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.140353 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.156926 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbmh5\" (UniqueName: \"kubernetes.io/projected/b683a409-b7d3-404f-ac5a-43be34e071c6-kube-api-access-mbmh5\") pod \"dnsmasq-dns-77585f5f8c-nv8xl\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.194171 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.624948 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-nv8xl"] Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.773680 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a845349-9de2-48bf-839d-cc06ec4a7619" path="/var/lib/kubelet/pods/1a845349-9de2-48bf-839d-cc06ec4a7619/volumes" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.843418 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-218e-account-create-l8rbx" Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.970619 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m98x5\" (UniqueName: \"kubernetes.io/projected/981cf93b-d455-4364-8d16-26b5d8e8ac56-kube-api-access-m98x5\") pod \"981cf93b-d455-4364-8d16-26b5d8e8ac56\" (UID: \"981cf93b-d455-4364-8d16-26b5d8e8ac56\") " Oct 10 13:32:56 crc kubenswrapper[4745]: I1010 13:32:56.974567 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/981cf93b-d455-4364-8d16-26b5d8e8ac56-kube-api-access-m98x5" (OuterVolumeSpecName: "kube-api-access-m98x5") pod "981cf93b-d455-4364-8d16-26b5d8e8ac56" (UID: "981cf93b-d455-4364-8d16-26b5d8e8ac56"). InnerVolumeSpecName "kube-api-access-m98x5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:32:57 crc kubenswrapper[4745]: I1010 13:32:57.072961 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m98x5\" (UniqueName: \"kubernetes.io/projected/981cf93b-d455-4364-8d16-26b5d8e8ac56-kube-api-access-m98x5\") on node \"crc\" DevicePath \"\"" Oct 10 13:32:57 crc kubenswrapper[4745]: I1010 13:32:57.584983 4745 generic.go:334] "Generic (PLEG): container finished" podID="b683a409-b7d3-404f-ac5a-43be34e071c6" containerID="4418d66aa17366f10075c640008b78b8e2f223f73a15dbd24ed2e4ace636096a" exitCode=0 Oct 10 13:32:57 crc kubenswrapper[4745]: I1010 13:32:57.585055 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" event={"ID":"b683a409-b7d3-404f-ac5a-43be34e071c6","Type":"ContainerDied","Data":"4418d66aa17366f10075c640008b78b8e2f223f73a15dbd24ed2e4ace636096a"} Oct 10 13:32:57 crc kubenswrapper[4745]: I1010 13:32:57.585103 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" event={"ID":"b683a409-b7d3-404f-ac5a-43be34e071c6","Type":"ContainerStarted","Data":"234d83317618d819612a8f20c6629adb2a1d793bf4c92b3dfcb3119df6d0e7d2"} Oct 10 13:32:57 crc kubenswrapper[4745]: I1010 13:32:57.586719 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-218e-account-create-l8rbx" event={"ID":"981cf93b-d455-4364-8d16-26b5d8e8ac56","Type":"ContainerDied","Data":"a6fb852bce7a8373ac30ac03ecd70cc89c65803b7e5ba8c0a6aa6ae1273fe84f"} Oct 10 13:32:57 crc kubenswrapper[4745]: I1010 13:32:57.586770 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6fb852bce7a8373ac30ac03ecd70cc89c65803b7e5ba8c0a6aa6ae1273fe84f" Oct 10 13:32:57 crc kubenswrapper[4745]: I1010 13:32:57.586773 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-218e-account-create-l8rbx" Oct 10 13:32:58 crc kubenswrapper[4745]: I1010 13:32:58.596718 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" event={"ID":"b683a409-b7d3-404f-ac5a-43be34e071c6","Type":"ContainerStarted","Data":"94760b0459376bca580d88cb72faa38ce3eb1269d08ffebd83ce22fcc54fb3ff"} Oct 10 13:32:58 crc kubenswrapper[4745]: I1010 13:32:58.597310 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:32:58 crc kubenswrapper[4745]: I1010 13:32:58.624320 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" podStartSLOduration=3.6242938049999998 podStartE2EDuration="3.624293805s" podCreationTimestamp="2025-10-10 13:32:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:32:58.62291141 +0000 UTC m=+892.520568203" watchObservedRunningTime="2025-10-10 13:32:58.624293805 +0000 UTC m=+892.521950578" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.163210 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-nfdck"] Oct 10 13:32:59 crc kubenswrapper[4745]: E1010 13:32:59.163714 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="981cf93b-d455-4364-8d16-26b5d8e8ac56" containerName="mariadb-account-create" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.163764 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="981cf93b-d455-4364-8d16-26b5d8e8ac56" containerName="mariadb-account-create" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.164035 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="981cf93b-d455-4364-8d16-26b5d8e8ac56" containerName="mariadb-account-create" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.164829 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-nfdck" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.167896 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.177198 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-nfdck"] Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.233031 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-x74z4" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.332951 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-config-data\") pod \"glance-db-sync-nfdck\" (UID: \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\") " pod="openstack/glance-db-sync-nfdck" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.333029 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-db-sync-config-data\") pod \"glance-db-sync-nfdck\" (UID: \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\") " pod="openstack/glance-db-sync-nfdck" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.333203 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-combined-ca-bundle\") pod \"glance-db-sync-nfdck\" (UID: \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\") " pod="openstack/glance-db-sync-nfdck" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.333281 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4qbr\" (UniqueName: \"kubernetes.io/projected/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-kube-api-access-g4qbr\") pod \"glance-db-sync-nfdck\" (UID: \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\") " pod="openstack/glance-db-sync-nfdck" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.434770 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-combined-ca-bundle\") pod \"glance-db-sync-nfdck\" (UID: \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\") " pod="openstack/glance-db-sync-nfdck" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.434818 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4qbr\" (UniqueName: \"kubernetes.io/projected/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-kube-api-access-g4qbr\") pod \"glance-db-sync-nfdck\" (UID: \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\") " pod="openstack/glance-db-sync-nfdck" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.434872 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-config-data\") pod \"glance-db-sync-nfdck\" (UID: \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\") " pod="openstack/glance-db-sync-nfdck" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.434929 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-db-sync-config-data\") pod \"glance-db-sync-nfdck\" (UID: \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\") " pod="openstack/glance-db-sync-nfdck" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.439175 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-db-sync-config-data\") pod \"glance-db-sync-nfdck\" (UID: \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\") " pod="openstack/glance-db-sync-nfdck" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.440847 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-combined-ca-bundle\") pod \"glance-db-sync-nfdck\" (UID: \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\") " pod="openstack/glance-db-sync-nfdck" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.442341 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-config-data\") pod \"glance-db-sync-nfdck\" (UID: \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\") " pod="openstack/glance-db-sync-nfdck" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.459159 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4qbr\" (UniqueName: \"kubernetes.io/projected/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-kube-api-access-g4qbr\") pod \"glance-db-sync-nfdck\" (UID: \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\") " pod="openstack/glance-db-sync-nfdck" Oct 10 13:32:59 crc kubenswrapper[4745]: I1010 13:32:59.542543 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-nfdck" Oct 10 13:33:00 crc kubenswrapper[4745]: I1010 13:33:00.036993 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-nfdck"] Oct 10 13:33:00 crc kubenswrapper[4745]: W1010 13:33:00.042023 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dfdede3_b1b6_4630_afe1_0c2cdecf57f1.slice/crio-6ed33c55a6f97844216b0939cd06b39297383a2d162e3434aafb06b428a59c3b WatchSource:0}: Error finding container 6ed33c55a6f97844216b0939cd06b39297383a2d162e3434aafb06b428a59c3b: Status 404 returned error can't find the container with id 6ed33c55a6f97844216b0939cd06b39297383a2d162e3434aafb06b428a59c3b Oct 10 13:33:00 crc kubenswrapper[4745]: I1010 13:33:00.615937 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-nfdck" event={"ID":"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1","Type":"ContainerStarted","Data":"6ed33c55a6f97844216b0939cd06b39297383a2d162e3434aafb06b428a59c3b"} Oct 10 13:33:04 crc kubenswrapper[4745]: I1010 13:33:04.965941 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.233855 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.323593 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-zw4m7"] Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.324683 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-zw4m7" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.370106 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-zw4m7"] Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.427263 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-7cqks"] Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.429655 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7cqks" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.433425 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-7cqks"] Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.454635 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw8qd\" (UniqueName: \"kubernetes.io/projected/75ea46f9-2e99-4f8d-ac69-053ba3f2e71d-kube-api-access-cw8qd\") pod \"cinder-db-create-zw4m7\" (UID: \"75ea46f9-2e99-4f8d-ac69-053ba3f2e71d\") " pod="openstack/cinder-db-create-zw4m7" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.515658 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-t5tfd"] Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.516674 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-t5tfd" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.524350 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-t5tfd"] Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.556481 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nkwb\" (UniqueName: \"kubernetes.io/projected/7c942887-aa59-417d-9ea5-e66cd98a1156-kube-api-access-4nkwb\") pod \"barbican-db-create-7cqks\" (UID: \"7c942887-aa59-417d-9ea5-e66cd98a1156\") " pod="openstack/barbican-db-create-7cqks" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.556598 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw8qd\" (UniqueName: \"kubernetes.io/projected/75ea46f9-2e99-4f8d-ac69-053ba3f2e71d-kube-api-access-cw8qd\") pod \"cinder-db-create-zw4m7\" (UID: \"75ea46f9-2e99-4f8d-ac69-053ba3f2e71d\") " pod="openstack/cinder-db-create-zw4m7" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.588594 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw8qd\" (UniqueName: \"kubernetes.io/projected/75ea46f9-2e99-4f8d-ac69-053ba3f2e71d-kube-api-access-cw8qd\") pod \"cinder-db-create-zw4m7\" (UID: \"75ea46f9-2e99-4f8d-ac69-053ba3f2e71d\") " pod="openstack/cinder-db-create-zw4m7" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.658424 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n987v\" (UniqueName: \"kubernetes.io/projected/531008bf-6675-4b72-83a5-3aaabf54a133-kube-api-access-n987v\") pod \"neutron-db-create-t5tfd\" (UID: \"531008bf-6675-4b72-83a5-3aaabf54a133\") " pod="openstack/neutron-db-create-t5tfd" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.658485 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-bdv5b"] Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.658498 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nkwb\" (UniqueName: \"kubernetes.io/projected/7c942887-aa59-417d-9ea5-e66cd98a1156-kube-api-access-4nkwb\") pod \"barbican-db-create-7cqks\" (UID: \"7c942887-aa59-417d-9ea5-e66cd98a1156\") " pod="openstack/barbican-db-create-7cqks" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.662913 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bdv5b" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.665864 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.666823 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.670458 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r7tdf" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.670481 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.672215 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-zw4m7" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.672765 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-bdv5b"] Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.685409 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nkwb\" (UniqueName: \"kubernetes.io/projected/7c942887-aa59-417d-9ea5-e66cd98a1156-kube-api-access-4nkwb\") pod \"barbican-db-create-7cqks\" (UID: \"7c942887-aa59-417d-9ea5-e66cd98a1156\") " pod="openstack/barbican-db-create-7cqks" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.757069 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7cqks" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.760692 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n987v\" (UniqueName: \"kubernetes.io/projected/531008bf-6675-4b72-83a5-3aaabf54a133-kube-api-access-n987v\") pod \"neutron-db-create-t5tfd\" (UID: \"531008bf-6675-4b72-83a5-3aaabf54a133\") " pod="openstack/neutron-db-create-t5tfd" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.760773 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d439664-95e3-4c4d-b22b-a14242767a3a-combined-ca-bundle\") pod \"keystone-db-sync-bdv5b\" (UID: \"2d439664-95e3-4c4d-b22b-a14242767a3a\") " pod="openstack/keystone-db-sync-bdv5b" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.760798 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d439664-95e3-4c4d-b22b-a14242767a3a-config-data\") pod \"keystone-db-sync-bdv5b\" (UID: \"2d439664-95e3-4c4d-b22b-a14242767a3a\") " pod="openstack/keystone-db-sync-bdv5b" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.760874 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nch9c\" (UniqueName: \"kubernetes.io/projected/2d439664-95e3-4c4d-b22b-a14242767a3a-kube-api-access-nch9c\") pod \"keystone-db-sync-bdv5b\" (UID: \"2d439664-95e3-4c4d-b22b-a14242767a3a\") " pod="openstack/keystone-db-sync-bdv5b" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.795747 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n987v\" (UniqueName: \"kubernetes.io/projected/531008bf-6675-4b72-83a5-3aaabf54a133-kube-api-access-n987v\") pod \"neutron-db-create-t5tfd\" (UID: \"531008bf-6675-4b72-83a5-3aaabf54a133\") " pod="openstack/neutron-db-create-t5tfd" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.862097 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d439664-95e3-4c4d-b22b-a14242767a3a-combined-ca-bundle\") pod \"keystone-db-sync-bdv5b\" (UID: \"2d439664-95e3-4c4d-b22b-a14242767a3a\") " pod="openstack/keystone-db-sync-bdv5b" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.862141 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d439664-95e3-4c4d-b22b-a14242767a3a-config-data\") pod \"keystone-db-sync-bdv5b\" (UID: \"2d439664-95e3-4c4d-b22b-a14242767a3a\") " pod="openstack/keystone-db-sync-bdv5b" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.862207 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nch9c\" (UniqueName: \"kubernetes.io/projected/2d439664-95e3-4c4d-b22b-a14242767a3a-kube-api-access-nch9c\") pod \"keystone-db-sync-bdv5b\" (UID: \"2d439664-95e3-4c4d-b22b-a14242767a3a\") " pod="openstack/keystone-db-sync-bdv5b" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.871375 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d439664-95e3-4c4d-b22b-a14242767a3a-combined-ca-bundle\") pod \"keystone-db-sync-bdv5b\" (UID: \"2d439664-95e3-4c4d-b22b-a14242767a3a\") " pod="openstack/keystone-db-sync-bdv5b" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.871778 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-t5tfd" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.875746 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d439664-95e3-4c4d-b22b-a14242767a3a-config-data\") pod \"keystone-db-sync-bdv5b\" (UID: \"2d439664-95e3-4c4d-b22b-a14242767a3a\") " pod="openstack/keystone-db-sync-bdv5b" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.883312 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nch9c\" (UniqueName: \"kubernetes.io/projected/2d439664-95e3-4c4d-b22b-a14242767a3a-kube-api-access-nch9c\") pod \"keystone-db-sync-bdv5b\" (UID: \"2d439664-95e3-4c4d-b22b-a14242767a3a\") " pod="openstack/keystone-db-sync-bdv5b" Oct 10 13:33:05 crc kubenswrapper[4745]: I1010 13:33:05.982130 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bdv5b" Oct 10 13:33:06 crc kubenswrapper[4745]: I1010 13:33:06.195578 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:33:06 crc kubenswrapper[4745]: I1010 13:33:06.256013 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2vx66"] Oct 10 13:33:06 crc kubenswrapper[4745]: I1010 13:33:06.256417 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-2vx66" podUID="ece46492-a015-4207-a04e-856390729ecb" containerName="dnsmasq-dns" containerID="cri-o://4598efad6dd2c1f942c5a23679903d92cd3c31b4618306670f3898369a166830" gracePeriod=10 Oct 10 13:33:06 crc kubenswrapper[4745]: I1010 13:33:06.673323 4745 generic.go:334] "Generic (PLEG): container finished" podID="ece46492-a015-4207-a04e-856390729ecb" containerID="4598efad6dd2c1f942c5a23679903d92cd3c31b4618306670f3898369a166830" exitCode=0 Oct 10 13:33:06 crc kubenswrapper[4745]: I1010 13:33:06.673367 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2vx66" event={"ID":"ece46492-a015-4207-a04e-856390729ecb","Type":"ContainerDied","Data":"4598efad6dd2c1f942c5a23679903d92cd3c31b4618306670f3898369a166830"} Oct 10 13:33:10 crc kubenswrapper[4745]: I1010 13:33:10.572445 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-2vx66" podUID="ece46492-a015-4207-a04e-856390729ecb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.130522 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.260433 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-ovsdbserver-sb\") pod \"ece46492-a015-4207-a04e-856390729ecb\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.260542 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-ovsdbserver-nb\") pod \"ece46492-a015-4207-a04e-856390729ecb\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.260665 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-config\") pod \"ece46492-a015-4207-a04e-856390729ecb\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.260688 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-dns-svc\") pod \"ece46492-a015-4207-a04e-856390729ecb\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.260715 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfp87\" (UniqueName: \"kubernetes.io/projected/ece46492-a015-4207-a04e-856390729ecb-kube-api-access-pfp87\") pod \"ece46492-a015-4207-a04e-856390729ecb\" (UID: \"ece46492-a015-4207-a04e-856390729ecb\") " Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.264771 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ece46492-a015-4207-a04e-856390729ecb-kube-api-access-pfp87" (OuterVolumeSpecName: "kube-api-access-pfp87") pod "ece46492-a015-4207-a04e-856390729ecb" (UID: "ece46492-a015-4207-a04e-856390729ecb"). InnerVolumeSpecName "kube-api-access-pfp87". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.347380 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ece46492-a015-4207-a04e-856390729ecb" (UID: "ece46492-a015-4207-a04e-856390729ecb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.348788 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ece46492-a015-4207-a04e-856390729ecb" (UID: "ece46492-a015-4207-a04e-856390729ecb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.360596 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-config" (OuterVolumeSpecName: "config") pod "ece46492-a015-4207-a04e-856390729ecb" (UID: "ece46492-a015-4207-a04e-856390729ecb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.362201 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.362313 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfp87\" (UniqueName: \"kubernetes.io/projected/ece46492-a015-4207-a04e-856390729ecb-kube-api-access-pfp87\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.362391 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.362451 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.372277 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ece46492-a015-4207-a04e-856390729ecb" (UID: "ece46492-a015-4207-a04e-856390729ecb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.467626 4745 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ece46492-a015-4207-a04e-856390729ecb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.473449 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-7cqks"] Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.481180 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-bdv5b"] Oct 10 13:33:11 crc kubenswrapper[4745]: W1010 13:33:11.486512 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d439664_95e3_4c4d_b22b_a14242767a3a.slice/crio-8c1bb473072e7eff642ef31a6b99cda37e25d39df1be73c4229336b97192ce8e WatchSource:0}: Error finding container 8c1bb473072e7eff642ef31a6b99cda37e25d39df1be73c4229336b97192ce8e: Status 404 returned error can't find the container with id 8c1bb473072e7eff642ef31a6b99cda37e25d39df1be73c4229336b97192ce8e Oct 10 13:33:11 crc kubenswrapper[4745]: W1010 13:33:11.488271 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c942887_aa59_417d_9ea5_e66cd98a1156.slice/crio-a1cce7dbaba0b8fde954fd47f5bbe4ba1ffe0c93fe114b4cf54cc3e1fd332d25 WatchSource:0}: Error finding container a1cce7dbaba0b8fde954fd47f5bbe4ba1ffe0c93fe114b4cf54cc3e1fd332d25: Status 404 returned error can't find the container with id a1cce7dbaba0b8fde954fd47f5bbe4ba1ffe0c93fe114b4cf54cc3e1fd332d25 Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.641718 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-zw4m7"] Oct 10 13:33:11 crc kubenswrapper[4745]: W1010 13:33:11.645843 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75ea46f9_2e99_4f8d_ac69_053ba3f2e71d.slice/crio-01c5094bbfd7c6533e146ff0c7eec54967eef9d2e7829af066285800c4619d58 WatchSource:0}: Error finding container 01c5094bbfd7c6533e146ff0c7eec54967eef9d2e7829af066285800c4619d58: Status 404 returned error can't find the container with id 01c5094bbfd7c6533e146ff0c7eec54967eef9d2e7829af066285800c4619d58 Oct 10 13:33:11 crc kubenswrapper[4745]: W1010 13:33:11.649411 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod531008bf_6675_4b72_83a5_3aaabf54a133.slice/crio-8c810d9bf185f1ccca82512a0c43fc7c7666216cd796aa1e7eef9b212369cc3b WatchSource:0}: Error finding container 8c810d9bf185f1ccca82512a0c43fc7c7666216cd796aa1e7eef9b212369cc3b: Status 404 returned error can't find the container with id 8c810d9bf185f1ccca82512a0c43fc7c7666216cd796aa1e7eef9b212369cc3b Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.652302 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-t5tfd"] Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.722825 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-t5tfd" event={"ID":"531008bf-6675-4b72-83a5-3aaabf54a133","Type":"ContainerStarted","Data":"8c810d9bf185f1ccca82512a0c43fc7c7666216cd796aa1e7eef9b212369cc3b"} Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.724834 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-nfdck" event={"ID":"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1","Type":"ContainerStarted","Data":"b76bccde80d9fa6432d17a2c1afe41626d6403bf036a993e9072ce4023acd22b"} Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.731027 4745 generic.go:334] "Generic (PLEG): container finished" podID="7c942887-aa59-417d-9ea5-e66cd98a1156" containerID="50b424a7887fcd3389e8d086c4b8417714a799fcf3d1110d2ab5273a60a5ec18" exitCode=0 Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.731155 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7cqks" event={"ID":"7c942887-aa59-417d-9ea5-e66cd98a1156","Type":"ContainerDied","Data":"50b424a7887fcd3389e8d086c4b8417714a799fcf3d1110d2ab5273a60a5ec18"} Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.731201 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7cqks" event={"ID":"7c942887-aa59-417d-9ea5-e66cd98a1156","Type":"ContainerStarted","Data":"a1cce7dbaba0b8fde954fd47f5bbe4ba1ffe0c93fe114b4cf54cc3e1fd332d25"} Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.735271 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-2vx66" Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.737200 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2vx66" event={"ID":"ece46492-a015-4207-a04e-856390729ecb","Type":"ContainerDied","Data":"7c695ed56ce50e5f7407f1113f70ad6c3677d5b9d6b8f23231646d1e1fc9e6e0"} Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.737318 4745 scope.go:117] "RemoveContainer" containerID="4598efad6dd2c1f942c5a23679903d92cd3c31b4618306670f3898369a166830" Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.741515 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bdv5b" event={"ID":"2d439664-95e3-4c4d-b22b-a14242767a3a","Type":"ContainerStarted","Data":"8c1bb473072e7eff642ef31a6b99cda37e25d39df1be73c4229336b97192ce8e"} Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.749560 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-zw4m7" event={"ID":"75ea46f9-2e99-4f8d-ac69-053ba3f2e71d","Type":"ContainerStarted","Data":"01c5094bbfd7c6533e146ff0c7eec54967eef9d2e7829af066285800c4619d58"} Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.758621 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-nfdck" podStartSLOduration=1.869114588 podStartE2EDuration="12.758593368s" podCreationTimestamp="2025-10-10 13:32:59 +0000 UTC" firstStartedPulling="2025-10-10 13:33:00.043596769 +0000 UTC m=+893.941253552" lastFinishedPulling="2025-10-10 13:33:10.933075529 +0000 UTC m=+904.830732332" observedRunningTime="2025-10-10 13:33:11.744129153 +0000 UTC m=+905.641785946" watchObservedRunningTime="2025-10-10 13:33:11.758593368 +0000 UTC m=+905.656250171" Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.777821 4745 scope.go:117] "RemoveContainer" containerID="4b53735422140493a15987542129134325505b81d926205f1d367d29361f88a5" Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.799858 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2vx66"] Oct 10 13:33:11 crc kubenswrapper[4745]: I1010 13:33:11.804642 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2vx66"] Oct 10 13:33:12 crc kubenswrapper[4745]: I1010 13:33:12.763754 4745 generic.go:334] "Generic (PLEG): container finished" podID="75ea46f9-2e99-4f8d-ac69-053ba3f2e71d" containerID="45effcc901700734d808a25087a434d0114d0bb624d3707963710d7581f4b9f9" exitCode=0 Oct 10 13:33:12 crc kubenswrapper[4745]: I1010 13:33:12.765556 4745 generic.go:334] "Generic (PLEG): container finished" podID="531008bf-6675-4b72-83a5-3aaabf54a133" containerID="d8e4e4690c3e55f3e93c6501f969fa2a4ea9030d0515f2fb876ab1921d0fb1e4" exitCode=0 Oct 10 13:33:12 crc kubenswrapper[4745]: I1010 13:33:12.767086 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ece46492-a015-4207-a04e-856390729ecb" path="/var/lib/kubelet/pods/ece46492-a015-4207-a04e-856390729ecb/volumes" Oct 10 13:33:12 crc kubenswrapper[4745]: I1010 13:33:12.769990 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-zw4m7" event={"ID":"75ea46f9-2e99-4f8d-ac69-053ba3f2e71d","Type":"ContainerDied","Data":"45effcc901700734d808a25087a434d0114d0bb624d3707963710d7581f4b9f9"} Oct 10 13:33:12 crc kubenswrapper[4745]: I1010 13:33:12.770040 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-t5tfd" event={"ID":"531008bf-6675-4b72-83a5-3aaabf54a133","Type":"ContainerDied","Data":"d8e4e4690c3e55f3e93c6501f969fa2a4ea9030d0515f2fb876ab1921d0fb1e4"} Oct 10 13:33:13 crc kubenswrapper[4745]: I1010 13:33:13.087661 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7cqks" Oct 10 13:33:13 crc kubenswrapper[4745]: I1010 13:33:13.200226 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nkwb\" (UniqueName: \"kubernetes.io/projected/7c942887-aa59-417d-9ea5-e66cd98a1156-kube-api-access-4nkwb\") pod \"7c942887-aa59-417d-9ea5-e66cd98a1156\" (UID: \"7c942887-aa59-417d-9ea5-e66cd98a1156\") " Oct 10 13:33:13 crc kubenswrapper[4745]: I1010 13:33:13.211881 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c942887-aa59-417d-9ea5-e66cd98a1156-kube-api-access-4nkwb" (OuterVolumeSpecName: "kube-api-access-4nkwb") pod "7c942887-aa59-417d-9ea5-e66cd98a1156" (UID: "7c942887-aa59-417d-9ea5-e66cd98a1156"). InnerVolumeSpecName "kube-api-access-4nkwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:33:13 crc kubenswrapper[4745]: I1010 13:33:13.303694 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nkwb\" (UniqueName: \"kubernetes.io/projected/7c942887-aa59-417d-9ea5-e66cd98a1156-kube-api-access-4nkwb\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:13 crc kubenswrapper[4745]: I1010 13:33:13.774214 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7cqks" Oct 10 13:33:13 crc kubenswrapper[4745]: I1010 13:33:13.782799 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7cqks" event={"ID":"7c942887-aa59-417d-9ea5-e66cd98a1156","Type":"ContainerDied","Data":"a1cce7dbaba0b8fde954fd47f5bbe4ba1ffe0c93fe114b4cf54cc3e1fd332d25"} Oct 10 13:33:13 crc kubenswrapper[4745]: I1010 13:33:13.782829 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1cce7dbaba0b8fde954fd47f5bbe4ba1ffe0c93fe114b4cf54cc3e1fd332d25" Oct 10 13:33:15 crc kubenswrapper[4745]: I1010 13:33:15.797315 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-zw4m7" event={"ID":"75ea46f9-2e99-4f8d-ac69-053ba3f2e71d","Type":"ContainerDied","Data":"01c5094bbfd7c6533e146ff0c7eec54967eef9d2e7829af066285800c4619d58"} Oct 10 13:33:15 crc kubenswrapper[4745]: I1010 13:33:15.797841 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01c5094bbfd7c6533e146ff0c7eec54967eef9d2e7829af066285800c4619d58" Oct 10 13:33:15 crc kubenswrapper[4745]: I1010 13:33:15.800072 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-t5tfd" event={"ID":"531008bf-6675-4b72-83a5-3aaabf54a133","Type":"ContainerDied","Data":"8c810d9bf185f1ccca82512a0c43fc7c7666216cd796aa1e7eef9b212369cc3b"} Oct 10 13:33:15 crc kubenswrapper[4745]: I1010 13:33:15.800103 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c810d9bf185f1ccca82512a0c43fc7c7666216cd796aa1e7eef9b212369cc3b" Oct 10 13:33:15 crc kubenswrapper[4745]: I1010 13:33:15.801421 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-zw4m7" Oct 10 13:33:15 crc kubenswrapper[4745]: I1010 13:33:15.947910 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cw8qd\" (UniqueName: \"kubernetes.io/projected/75ea46f9-2e99-4f8d-ac69-053ba3f2e71d-kube-api-access-cw8qd\") pod \"75ea46f9-2e99-4f8d-ac69-053ba3f2e71d\" (UID: \"75ea46f9-2e99-4f8d-ac69-053ba3f2e71d\") " Oct 10 13:33:15 crc kubenswrapper[4745]: I1010 13:33:15.951150 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75ea46f9-2e99-4f8d-ac69-053ba3f2e71d-kube-api-access-cw8qd" (OuterVolumeSpecName: "kube-api-access-cw8qd") pod "75ea46f9-2e99-4f8d-ac69-053ba3f2e71d" (UID: "75ea46f9-2e99-4f8d-ac69-053ba3f2e71d"). InnerVolumeSpecName "kube-api-access-cw8qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:33:15 crc kubenswrapper[4745]: I1010 13:33:15.953993 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-t5tfd" Oct 10 13:33:16 crc kubenswrapper[4745]: I1010 13:33:16.050317 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cw8qd\" (UniqueName: \"kubernetes.io/projected/75ea46f9-2e99-4f8d-ac69-053ba3f2e71d-kube-api-access-cw8qd\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:16 crc kubenswrapper[4745]: I1010 13:33:16.151293 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n987v\" (UniqueName: \"kubernetes.io/projected/531008bf-6675-4b72-83a5-3aaabf54a133-kube-api-access-n987v\") pod \"531008bf-6675-4b72-83a5-3aaabf54a133\" (UID: \"531008bf-6675-4b72-83a5-3aaabf54a133\") " Oct 10 13:33:16 crc kubenswrapper[4745]: I1010 13:33:16.155006 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/531008bf-6675-4b72-83a5-3aaabf54a133-kube-api-access-n987v" (OuterVolumeSpecName: "kube-api-access-n987v") pod "531008bf-6675-4b72-83a5-3aaabf54a133" (UID: "531008bf-6675-4b72-83a5-3aaabf54a133"). InnerVolumeSpecName "kube-api-access-n987v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:33:16 crc kubenswrapper[4745]: I1010 13:33:16.187566 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:33:16 crc kubenswrapper[4745]: I1010 13:33:16.187660 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:33:16 crc kubenswrapper[4745]: I1010 13:33:16.253633 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n987v\" (UniqueName: \"kubernetes.io/projected/531008bf-6675-4b72-83a5-3aaabf54a133-kube-api-access-n987v\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:16 crc kubenswrapper[4745]: I1010 13:33:16.808379 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bdv5b" event={"ID":"2d439664-95e3-4c4d-b22b-a14242767a3a","Type":"ContainerStarted","Data":"45e243e61dc30b54e185fd498af873db27c34aba4f03cd1ae9945b80fe22815a"} Oct 10 13:33:16 crc kubenswrapper[4745]: I1010 13:33:16.808412 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-t5tfd" Oct 10 13:33:16 crc kubenswrapper[4745]: I1010 13:33:16.811362 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-zw4m7" Oct 10 13:33:16 crc kubenswrapper[4745]: I1010 13:33:16.834507 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-bdv5b" podStartSLOduration=7.533778107 podStartE2EDuration="11.834492269s" podCreationTimestamp="2025-10-10 13:33:05 +0000 UTC" firstStartedPulling="2025-10-10 13:33:11.488901456 +0000 UTC m=+905.386558219" lastFinishedPulling="2025-10-10 13:33:15.789615608 +0000 UTC m=+909.687272381" observedRunningTime="2025-10-10 13:33:16.829112373 +0000 UTC m=+910.726769206" watchObservedRunningTime="2025-10-10 13:33:16.834492269 +0000 UTC m=+910.732149032" Oct 10 13:33:17 crc kubenswrapper[4745]: I1010 13:33:17.826287 4745 generic.go:334] "Generic (PLEG): container finished" podID="6dfdede3-b1b6-4630-afe1-0c2cdecf57f1" containerID="b76bccde80d9fa6432d17a2c1afe41626d6403bf036a993e9072ce4023acd22b" exitCode=0 Oct 10 13:33:17 crc kubenswrapper[4745]: I1010 13:33:17.826407 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-nfdck" event={"ID":"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1","Type":"ContainerDied","Data":"b76bccde80d9fa6432d17a2c1afe41626d6403bf036a993e9072ce4023acd22b"} Oct 10 13:33:18 crc kubenswrapper[4745]: I1010 13:33:18.838528 4745 generic.go:334] "Generic (PLEG): container finished" podID="2d439664-95e3-4c4d-b22b-a14242767a3a" containerID="45e243e61dc30b54e185fd498af873db27c34aba4f03cd1ae9945b80fe22815a" exitCode=0 Oct 10 13:33:18 crc kubenswrapper[4745]: I1010 13:33:18.838642 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bdv5b" event={"ID":"2d439664-95e3-4c4d-b22b-a14242767a3a","Type":"ContainerDied","Data":"45e243e61dc30b54e185fd498af873db27c34aba4f03cd1ae9945b80fe22815a"} Oct 10 13:33:19 crc kubenswrapper[4745]: I1010 13:33:19.314154 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-nfdck" Oct 10 13:33:19 crc kubenswrapper[4745]: I1010 13:33:19.411782 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-config-data\") pod \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\" (UID: \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\") " Oct 10 13:33:19 crc kubenswrapper[4745]: I1010 13:33:19.411895 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-db-sync-config-data\") pod \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\" (UID: \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\") " Oct 10 13:33:19 crc kubenswrapper[4745]: I1010 13:33:19.412004 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-combined-ca-bundle\") pod \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\" (UID: \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\") " Oct 10 13:33:19 crc kubenswrapper[4745]: I1010 13:33:19.412050 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4qbr\" (UniqueName: \"kubernetes.io/projected/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-kube-api-access-g4qbr\") pod \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\" (UID: \"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1\") " Oct 10 13:33:19 crc kubenswrapper[4745]: I1010 13:33:19.416968 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6dfdede3-b1b6-4630-afe1-0c2cdecf57f1" (UID: "6dfdede3-b1b6-4630-afe1-0c2cdecf57f1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:19 crc kubenswrapper[4745]: I1010 13:33:19.419573 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-kube-api-access-g4qbr" (OuterVolumeSpecName: "kube-api-access-g4qbr") pod "6dfdede3-b1b6-4630-afe1-0c2cdecf57f1" (UID: "6dfdede3-b1b6-4630-afe1-0c2cdecf57f1"). InnerVolumeSpecName "kube-api-access-g4qbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:33:19 crc kubenswrapper[4745]: I1010 13:33:19.438171 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6dfdede3-b1b6-4630-afe1-0c2cdecf57f1" (UID: "6dfdede3-b1b6-4630-afe1-0c2cdecf57f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:19 crc kubenswrapper[4745]: I1010 13:33:19.459954 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-config-data" (OuterVolumeSpecName: "config-data") pod "6dfdede3-b1b6-4630-afe1-0c2cdecf57f1" (UID: "6dfdede3-b1b6-4630-afe1-0c2cdecf57f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:19 crc kubenswrapper[4745]: I1010 13:33:19.514009 4745 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:19 crc kubenswrapper[4745]: I1010 13:33:19.514038 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:19 crc kubenswrapper[4745]: I1010 13:33:19.514048 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4qbr\" (UniqueName: \"kubernetes.io/projected/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-kube-api-access-g4qbr\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:19 crc kubenswrapper[4745]: I1010 13:33:19.514059 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:19 crc kubenswrapper[4745]: I1010 13:33:19.852074 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-nfdck" event={"ID":"6dfdede3-b1b6-4630-afe1-0c2cdecf57f1","Type":"ContainerDied","Data":"6ed33c55a6f97844216b0939cd06b39297383a2d162e3434aafb06b428a59c3b"} Oct 10 13:33:19 crc kubenswrapper[4745]: I1010 13:33:19.852442 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ed33c55a6f97844216b0939cd06b39297383a2d162e3434aafb06b428a59c3b" Oct 10 13:33:19 crc kubenswrapper[4745]: I1010 13:33:19.852145 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-nfdck" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.129208 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bdv5b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.225918 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d439664-95e3-4c4d-b22b-a14242767a3a-config-data\") pod \"2d439664-95e3-4c4d-b22b-a14242767a3a\" (UID: \"2d439664-95e3-4c4d-b22b-a14242767a3a\") " Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.226026 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nch9c\" (UniqueName: \"kubernetes.io/projected/2d439664-95e3-4c4d-b22b-a14242767a3a-kube-api-access-nch9c\") pod \"2d439664-95e3-4c4d-b22b-a14242767a3a\" (UID: \"2d439664-95e3-4c4d-b22b-a14242767a3a\") " Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.226064 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d439664-95e3-4c4d-b22b-a14242767a3a-combined-ca-bundle\") pod \"2d439664-95e3-4c4d-b22b-a14242767a3a\" (UID: \"2d439664-95e3-4c4d-b22b-a14242767a3a\") " Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.234280 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d439664-95e3-4c4d-b22b-a14242767a3a-kube-api-access-nch9c" (OuterVolumeSpecName: "kube-api-access-nch9c") pod "2d439664-95e3-4c4d-b22b-a14242767a3a" (UID: "2d439664-95e3-4c4d-b22b-a14242767a3a"). InnerVolumeSpecName "kube-api-access-nch9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.295903 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-cw95b"] Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.298360 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d439664-95e3-4c4d-b22b-a14242767a3a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d439664-95e3-4c4d-b22b-a14242767a3a" (UID: "2d439664-95e3-4c4d-b22b-a14242767a3a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:20 crc kubenswrapper[4745]: E1010 13:33:20.301051 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ece46492-a015-4207-a04e-856390729ecb" containerName="dnsmasq-dns" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.311165 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="ece46492-a015-4207-a04e-856390729ecb" containerName="dnsmasq-dns" Oct 10 13:33:20 crc kubenswrapper[4745]: E1010 13:33:20.311271 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="531008bf-6675-4b72-83a5-3aaabf54a133" containerName="mariadb-database-create" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.311296 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="531008bf-6675-4b72-83a5-3aaabf54a133" containerName="mariadb-database-create" Oct 10 13:33:20 crc kubenswrapper[4745]: E1010 13:33:20.311308 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d439664-95e3-4c4d-b22b-a14242767a3a" containerName="keystone-db-sync" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.311315 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d439664-95e3-4c4d-b22b-a14242767a3a" containerName="keystone-db-sync" Oct 10 13:33:20 crc kubenswrapper[4745]: E1010 13:33:20.311350 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dfdede3-b1b6-4630-afe1-0c2cdecf57f1" containerName="glance-db-sync" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.311356 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dfdede3-b1b6-4630-afe1-0c2cdecf57f1" containerName="glance-db-sync" Oct 10 13:33:20 crc kubenswrapper[4745]: E1010 13:33:20.311365 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75ea46f9-2e99-4f8d-ac69-053ba3f2e71d" containerName="mariadb-database-create" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.311371 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="75ea46f9-2e99-4f8d-ac69-053ba3f2e71d" containerName="mariadb-database-create" Oct 10 13:33:20 crc kubenswrapper[4745]: E1010 13:33:20.311383 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c942887-aa59-417d-9ea5-e66cd98a1156" containerName="mariadb-database-create" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.311390 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c942887-aa59-417d-9ea5-e66cd98a1156" containerName="mariadb-database-create" Oct 10 13:33:20 crc kubenswrapper[4745]: E1010 13:33:20.311400 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ece46492-a015-4207-a04e-856390729ecb" containerName="init" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.311421 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="ece46492-a015-4207-a04e-856390729ecb" containerName="init" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.313161 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="ece46492-a015-4207-a04e-856390729ecb" containerName="dnsmasq-dns" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.313188 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d439664-95e3-4c4d-b22b-a14242767a3a" containerName="keystone-db-sync" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.313199 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="531008bf-6675-4b72-83a5-3aaabf54a133" containerName="mariadb-database-create" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.313212 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dfdede3-b1b6-4630-afe1-0c2cdecf57f1" containerName="glance-db-sync" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.313238 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c942887-aa59-417d-9ea5-e66cd98a1156" containerName="mariadb-database-create" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.313245 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="75ea46f9-2e99-4f8d-ac69-053ba3f2e71d" containerName="mariadb-database-create" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.314214 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.314268 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-cw95b"] Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.322312 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d439664-95e3-4c4d-b22b-a14242767a3a-config-data" (OuterVolumeSpecName: "config-data") pod "2d439664-95e3-4c4d-b22b-a14242767a3a" (UID: "2d439664-95e3-4c4d-b22b-a14242767a3a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.335799 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d439664-95e3-4c4d-b22b-a14242767a3a-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.336417 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nch9c\" (UniqueName: \"kubernetes.io/projected/2d439664-95e3-4c4d-b22b-a14242767a3a-kube-api-access-nch9c\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.336489 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d439664-95e3-4c4d-b22b-a14242767a3a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.440553 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.440621 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-config\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.440643 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.440679 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5fvc\" (UniqueName: \"kubernetes.io/projected/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-kube-api-access-h5fvc\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.440707 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.440739 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.541832 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.541906 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-config\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.541925 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.541966 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5fvc\" (UniqueName: \"kubernetes.io/projected/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-kube-api-access-h5fvc\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.541994 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.542020 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.542916 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.542941 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.543217 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-config\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.543532 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.543891 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.566898 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5fvc\" (UniqueName: \"kubernetes.io/projected/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-kube-api-access-h5fvc\") pod \"dnsmasq-dns-7ff5475cc9-cw95b\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.726052 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.863297 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bdv5b" event={"ID":"2d439664-95e3-4c4d-b22b-a14242767a3a","Type":"ContainerDied","Data":"8c1bb473072e7eff642ef31a6b99cda37e25d39df1be73c4229336b97192ce8e"} Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.863536 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c1bb473072e7eff642ef31a6b99cda37e25d39df1be73c4229336b97192ce8e" Oct 10 13:33:20 crc kubenswrapper[4745]: I1010 13:33:20.863583 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bdv5b" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.063841 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-cw95b"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.071958 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-nwttm"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.073054 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.076769 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.076978 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.077114 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.077315 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r7tdf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.103935 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nwttm"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.119122 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.120385 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.168595 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-config-data\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.168904 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q9rg\" (UniqueName: \"kubernetes.io/projected/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-kube-api-access-6q9rg\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.169115 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-fernet-keys\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.169216 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-combined-ca-bundle\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.169369 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-credential-keys\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.169467 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-scripts\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.192810 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.221797 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6fcdc9c87c-9r2nf"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.223205 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.225411 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.225606 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-g5hkn" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.225778 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.225887 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.271020 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6fcdc9c87c-9r2nf"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.272252 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxtk4\" (UniqueName: \"kubernetes.io/projected/581b093c-c14e-499e-b5da-5591754ad44a-kube-api-access-pxtk4\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.272295 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-credential-keys\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.272319 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-scripts\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.272342 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.272376 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-config-data\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.272399 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.272427 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q9rg\" (UniqueName: \"kubernetes.io/projected/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-kube-api-access-6q9rg\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.272456 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-config\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.272488 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.272523 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.272542 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-fernet-keys\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.272569 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-combined-ca-bundle\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.278368 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-credential-keys\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.282330 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-fernet-keys\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.282490 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-cw95b"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.301560 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-combined-ca-bundle\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.302168 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-config-data\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.305862 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-scripts\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.307564 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q9rg\" (UniqueName: \"kubernetes.io/projected/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-kube-api-access-6q9rg\") pod \"keystone-bootstrap-nwttm\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.376471 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/83abb158-a66f-48ed-ae55-40a2c35cfbf2-horizon-secret-key\") pod \"horizon-6fcdc9c87c-9r2nf\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.376527 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.376568 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxtk4\" (UniqueName: \"kubernetes.io/projected/581b093c-c14e-499e-b5da-5591754ad44a-kube-api-access-pxtk4\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.376583 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83abb158-a66f-48ed-ae55-40a2c35cfbf2-logs\") pod \"horizon-6fcdc9c87c-9r2nf\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.376619 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.376645 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83abb158-a66f-48ed-ae55-40a2c35cfbf2-scripts\") pod \"horizon-6fcdc9c87c-9r2nf\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.376687 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.376706 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/83abb158-a66f-48ed-ae55-40a2c35cfbf2-config-data\") pod \"horizon-6fcdc9c87c-9r2nf\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.376762 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-config\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.376792 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.376807 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gv94\" (UniqueName: \"kubernetes.io/projected/83abb158-a66f-48ed-ae55-40a2c35cfbf2-kube-api-access-2gv94\") pod \"horizon-6fcdc9c87c-9r2nf\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.377582 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.378415 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.378950 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.379460 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-config\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.379985 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.391483 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.418579 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxtk4\" (UniqueName: \"kubernetes.io/projected/581b093c-c14e-499e-b5da-5591754ad44a-kube-api-access-pxtk4\") pod \"dnsmasq-dns-5c5cc7c5ff-fn2x8\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.469074 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-66b896bb7f-64vcw"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.470527 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.482284 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83abb158-a66f-48ed-ae55-40a2c35cfbf2-logs\") pod \"horizon-6fcdc9c87c-9r2nf\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.482333 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83abb158-a66f-48ed-ae55-40a2c35cfbf2-scripts\") pod \"horizon-6fcdc9c87c-9r2nf\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.482369 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/83abb158-a66f-48ed-ae55-40a2c35cfbf2-config-data\") pod \"horizon-6fcdc9c87c-9r2nf\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.482427 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gv94\" (UniqueName: \"kubernetes.io/projected/83abb158-a66f-48ed-ae55-40a2c35cfbf2-kube-api-access-2gv94\") pod \"horizon-6fcdc9c87c-9r2nf\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.482450 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/83abb158-a66f-48ed-ae55-40a2c35cfbf2-horizon-secret-key\") pod \"horizon-6fcdc9c87c-9r2nf\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.483633 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83abb158-a66f-48ed-ae55-40a2c35cfbf2-scripts\") pod \"horizon-6fcdc9c87c-9r2nf\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.483898 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83abb158-a66f-48ed-ae55-40a2c35cfbf2-logs\") pod \"horizon-6fcdc9c87c-9r2nf\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.484765 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/83abb158-a66f-48ed-ae55-40a2c35cfbf2-config-data\") pod \"horizon-6fcdc9c87c-9r2nf\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.490324 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/83abb158-a66f-48ed-ae55-40a2c35cfbf2-horizon-secret-key\") pod \"horizon-6fcdc9c87c-9r2nf\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.496578 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.506791 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66b896bb7f-64vcw"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.515591 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.517596 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.520171 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.520353 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.525111 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.526558 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.530565 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.537151 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.537410 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-x74z4" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.537570 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.538939 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.554571 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gv94\" (UniqueName: \"kubernetes.io/projected/83abb158-a66f-48ed-ae55-40a2c35cfbf2-kube-api-access-2gv94\") pod \"horizon-6fcdc9c87c-9r2nf\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.554627 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.564809 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-c2vrp"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.566688 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-c2vrp" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.580654 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-9dnj8" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.580693 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.580827 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.586609 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj2lz\" (UniqueName: \"kubernetes.io/projected/17718d89-5706-452c-a482-027b2c6f7c0d-kube-api-access-mj2lz\") pod \"horizon-66b896bb7f-64vcw\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.586653 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17718d89-5706-452c-a482-027b2c6f7c0d-logs\") pod \"horizon-66b896bb7f-64vcw\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.586692 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17718d89-5706-452c-a482-027b2c6f7c0d-config-data\") pod \"horizon-66b896bb7f-64vcw\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.586723 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17718d89-5706-452c-a482-027b2c6f7c0d-scripts\") pod \"horizon-66b896bb7f-64vcw\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.586760 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17718d89-5706-452c-a482-027b2c6f7c0d-horizon-secret-key\") pod \"horizon-66b896bb7f-64vcw\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.602047 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-c2vrp"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.639705 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-vhj7g"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.641047 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.652968 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-vhj7g"] Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.687631 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-scripts\") pod \"placement-db-sync-c2vrp\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " pod="openstack/placement-db-sync-c2vrp" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.687997 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w9cx\" (UniqueName: \"kubernetes.io/projected/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-kube-api-access-5w9cx\") pod \"placement-db-sync-c2vrp\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " pod="openstack/placement-db-sync-c2vrp" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688020 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688042 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5e0c23f6-8316-4e64-8d05-02d984fdb66e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688066 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj2lz\" (UniqueName: \"kubernetes.io/projected/17718d89-5706-452c-a482-027b2c6f7c0d-kube-api-access-mj2lz\") pod \"horizon-66b896bb7f-64vcw\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688082 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wz9r\" (UniqueName: \"kubernetes.io/projected/822d9514-5b19-440a-b4dc-5850f2bc8549-kube-api-access-2wz9r\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688106 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17718d89-5706-452c-a482-027b2c6f7c0d-logs\") pod \"horizon-66b896bb7f-64vcw\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688136 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/822d9514-5b19-440a-b4dc-5850f2bc8549-log-httpd\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688164 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-config-data\") pod \"placement-db-sync-c2vrp\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " pod="openstack/placement-db-sync-c2vrp" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688190 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17718d89-5706-452c-a482-027b2c6f7c0d-config-data\") pod \"horizon-66b896bb7f-64vcw\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688215 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j4bn\" (UniqueName: \"kubernetes.io/projected/5e0c23f6-8316-4e64-8d05-02d984fdb66e-kube-api-access-9j4bn\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688236 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-scripts\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688254 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/822d9514-5b19-440a-b4dc-5850f2bc8549-run-httpd\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688275 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17718d89-5706-452c-a482-027b2c6f7c0d-scripts\") pod \"horizon-66b896bb7f-64vcw\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688299 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17718d89-5706-452c-a482-027b2c6f7c0d-horizon-secret-key\") pod \"horizon-66b896bb7f-64vcw\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688326 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688344 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688360 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-logs\") pod \"placement-db-sync-c2vrp\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " pod="openstack/placement-db-sync-c2vrp" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688383 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-combined-ca-bundle\") pod \"placement-db-sync-c2vrp\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " pod="openstack/placement-db-sync-c2vrp" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688400 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-config-data\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688415 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-config-data\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688441 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e0c23f6-8316-4e64-8d05-02d984fdb66e-logs\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688464 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.688497 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-scripts\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.689130 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17718d89-5706-452c-a482-027b2c6f7c0d-logs\") pod \"horizon-66b896bb7f-64vcw\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.690809 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17718d89-5706-452c-a482-027b2c6f7c0d-config-data\") pod \"horizon-66b896bb7f-64vcw\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.691201 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17718d89-5706-452c-a482-027b2c6f7c0d-scripts\") pod \"horizon-66b896bb7f-64vcw\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.698030 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17718d89-5706-452c-a482-027b2c6f7c0d-horizon-secret-key\") pod \"horizon-66b896bb7f-64vcw\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.714469 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj2lz\" (UniqueName: \"kubernetes.io/projected/17718d89-5706-452c-a482-027b2c6f7c0d-kube-api-access-mj2lz\") pod \"horizon-66b896bb7f-64vcw\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.791279 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e0c23f6-8316-4e64-8d05-02d984fdb66e-logs\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.791338 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.791396 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-scripts\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.791431 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-config\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.791454 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f658\" (UniqueName: \"kubernetes.io/projected/2abd8e35-90b0-49c6-935e-2b757228070b-kube-api-access-9f658\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.791880 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-scripts\") pod \"placement-db-sync-c2vrp\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " pod="openstack/placement-db-sync-c2vrp" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.791895 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e0c23f6-8316-4e64-8d05-02d984fdb66e-logs\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.791825 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.796268 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-scripts\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.798280 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-scripts\") pod \"placement-db-sync-c2vrp\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " pod="openstack/placement-db-sync-c2vrp" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807014 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w9cx\" (UniqueName: \"kubernetes.io/projected/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-kube-api-access-5w9cx\") pod \"placement-db-sync-c2vrp\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " pod="openstack/placement-db-sync-c2vrp" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807057 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807084 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5e0c23f6-8316-4e64-8d05-02d984fdb66e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807115 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wz9r\" (UniqueName: \"kubernetes.io/projected/822d9514-5b19-440a-b4dc-5850f2bc8549-kube-api-access-2wz9r\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807135 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807194 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/822d9514-5b19-440a-b4dc-5850f2bc8549-log-httpd\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807228 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807251 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-config-data\") pod \"placement-db-sync-c2vrp\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " pod="openstack/placement-db-sync-c2vrp" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807287 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807318 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j4bn\" (UniqueName: \"kubernetes.io/projected/5e0c23f6-8316-4e64-8d05-02d984fdb66e-kube-api-access-9j4bn\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807341 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-scripts\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807361 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/822d9514-5b19-440a-b4dc-5850f2bc8549-run-httpd\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807421 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807449 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807464 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-logs\") pod \"placement-db-sync-c2vrp\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " pod="openstack/placement-db-sync-c2vrp" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807490 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-combined-ca-bundle\") pod \"placement-db-sync-c2vrp\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " pod="openstack/placement-db-sync-c2vrp" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807509 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-config-data\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807525 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-config-data\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.807553 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.808213 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5e0c23f6-8316-4e64-8d05-02d984fdb66e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.808322 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/822d9514-5b19-440a-b4dc-5850f2bc8549-log-httpd\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.808570 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-logs\") pod \"placement-db-sync-c2vrp\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " pod="openstack/placement-db-sync-c2vrp" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.809157 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/822d9514-5b19-440a-b4dc-5850f2bc8549-run-httpd\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.815341 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.821198 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.823495 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-config-data\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.824110 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-combined-ca-bundle\") pod \"placement-db-sync-c2vrp\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " pod="openstack/placement-db-sync-c2vrp" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.826390 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.829756 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-config-data\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.829794 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j4bn\" (UniqueName: \"kubernetes.io/projected/5e0c23f6-8316-4e64-8d05-02d984fdb66e-kube-api-access-9j4bn\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.830261 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wz9r\" (UniqueName: \"kubernetes.io/projected/822d9514-5b19-440a-b4dc-5850f2bc8549-kube-api-access-2wz9r\") pod \"ceilometer-0\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.830379 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w9cx\" (UniqueName: \"kubernetes.io/projected/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-kube-api-access-5w9cx\") pod \"placement-db-sync-c2vrp\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " pod="openstack/placement-db-sync-c2vrp" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.834766 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-config-data\") pod \"placement-db-sync-c2vrp\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " pod="openstack/placement-db-sync-c2vrp" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.853093 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.854241 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-scripts\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.854723 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.880632 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.895713 4745 generic.go:334] "Generic (PLEG): container finished" podID="7d87ab1d-2e84-448a-ad4e-9b672d6922a6" containerID="716e3b93ed72d18ec3c1b8c14604171e88d17bc03cdc9ee1e0c8325bf6fbd522" exitCode=0 Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.895786 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" event={"ID":"7d87ab1d-2e84-448a-ad4e-9b672d6922a6","Type":"ContainerDied","Data":"716e3b93ed72d18ec3c1b8c14604171e88d17bc03cdc9ee1e0c8325bf6fbd522"} Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.895809 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" event={"ID":"7d87ab1d-2e84-448a-ad4e-9b672d6922a6","Type":"ContainerStarted","Data":"495097d934f0637c5c429b5cc33065b851a4cdf31cc74ed9019d3ec583ef0d65"} Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.914553 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.914613 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.914685 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.914745 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-config\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.914764 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f658\" (UniqueName: \"kubernetes.io/projected/2abd8e35-90b0-49c6-935e-2b757228070b-kube-api-access-9f658\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.914792 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.916196 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.916658 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.917345 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.925959 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-config\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.926511 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.927968 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.977657 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 13:33:21 crc kubenswrapper[4745]: I1010 13:33:21.988603 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f658\" (UniqueName: \"kubernetes.io/projected/2abd8e35-90b0-49c6-935e-2b757228070b-kube-api-access-9f658\") pod \"dnsmasq-dns-8b5c85b87-vhj7g\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.018889 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.019399 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-c2vrp" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.083770 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nwttm"] Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.233976 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8"] Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.350539 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.352228 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.360361 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.370370 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.432227 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.432269 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.432311 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.432333 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122e3abe-be44-46cc-826f-938bb985e0b2-logs\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.432489 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.432518 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/122e3abe-be44-46cc-826f-938bb985e0b2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.432567 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgkcx\" (UniqueName: \"kubernetes.io/projected/122e3abe-be44-46cc-826f-938bb985e0b2-kube-api-access-mgkcx\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.533242 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.533285 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/122e3abe-be44-46cc-826f-938bb985e0b2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.533320 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgkcx\" (UniqueName: \"kubernetes.io/projected/122e3abe-be44-46cc-826f-938bb985e0b2-kube-api-access-mgkcx\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.533374 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.533394 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.533433 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.533452 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122e3abe-be44-46cc-826f-938bb985e0b2-logs\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.534124 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/122e3abe-be44-46cc-826f-938bb985e0b2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.534170 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122e3abe-be44-46cc-826f-938bb985e0b2-logs\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.534351 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.539588 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.542151 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.543255 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.556800 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgkcx\" (UniqueName: \"kubernetes.io/projected/122e3abe-be44-46cc-826f-938bb985e0b2-kube-api-access-mgkcx\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.576513 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.636650 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.642616 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.683709 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66b896bb7f-64vcw"] Oct 10 13:33:22 crc kubenswrapper[4745]: W1010 13:33:22.686248 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17718d89_5706_452c_a482_027b2c6f7c0d.slice/crio-6bc8a5904b5bb03d069ec0a58dc80f185db938ce6f84821a384b5d5b4aa771de WatchSource:0}: Error finding container 6bc8a5904b5bb03d069ec0a58dc80f185db938ce6f84821a384b5d5b4aa771de: Status 404 returned error can't find the container with id 6bc8a5904b5bb03d069ec0a58dc80f185db938ce6f84821a384b5d5b4aa771de Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.703717 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6fcdc9c87c-9r2nf"] Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.714145 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:33:22 crc kubenswrapper[4745]: W1010 13:33:22.734575 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod822d9514_5b19_440a_b4dc_5850f2bc8549.slice/crio-ae3bdb198f4bac50a10007e2b392a3a634e62adcd0eb931cf2cd264d3b9cb7e2 WatchSource:0}: Error finding container ae3bdb198f4bac50a10007e2b392a3a634e62adcd0eb931cf2cd264d3b9cb7e2: Status 404 returned error can't find the container with id ae3bdb198f4bac50a10007e2b392a3a634e62adcd0eb931cf2cd264d3b9cb7e2 Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.837931 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-config\") pod \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.838005 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-dns-svc\") pod \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.838042 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-dns-swift-storage-0\") pod \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.838129 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-ovsdbserver-sb\") pod \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.838226 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-ovsdbserver-nb\") pod \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.838255 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5fvc\" (UniqueName: \"kubernetes.io/projected/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-kube-api-access-h5fvc\") pod \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\" (UID: \"7d87ab1d-2e84-448a-ad4e-9b672d6922a6\") " Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.845358 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-kube-api-access-h5fvc" (OuterVolumeSpecName: "kube-api-access-h5fvc") pod "7d87ab1d-2e84-448a-ad4e-9b672d6922a6" (UID: "7d87ab1d-2e84-448a-ad4e-9b672d6922a6"). InnerVolumeSpecName "kube-api-access-h5fvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.856606 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-vhj7g"] Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.865442 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7d87ab1d-2e84-448a-ad4e-9b672d6922a6" (UID: "7d87ab1d-2e84-448a-ad4e-9b672d6922a6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.883129 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-config" (OuterVolumeSpecName: "config") pod "7d87ab1d-2e84-448a-ad4e-9b672d6922a6" (UID: "7d87ab1d-2e84-448a-ad4e-9b672d6922a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.895318 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-c2vrp"] Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.896956 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7d87ab1d-2e84-448a-ad4e-9b672d6922a6" (UID: "7d87ab1d-2e84-448a-ad4e-9b672d6922a6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.912533 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7d87ab1d-2e84-448a-ad4e-9b672d6922a6" (UID: "7d87ab1d-2e84-448a-ad4e-9b672d6922a6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.914310 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"822d9514-5b19-440a-b4dc-5850f2bc8549","Type":"ContainerStarted","Data":"ae3bdb198f4bac50a10007e2b392a3a634e62adcd0eb931cf2cd264d3b9cb7e2"} Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.916580 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fcdc9c87c-9r2nf" event={"ID":"83abb158-a66f-48ed-ae55-40a2c35cfbf2","Type":"ContainerStarted","Data":"5486e6e9d0a18e48d94f08e3c2f3f6dfb047e1ba2f47d1785d8b378fe3dc6de8"} Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.919355 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" event={"ID":"7d87ab1d-2e84-448a-ad4e-9b672d6922a6","Type":"ContainerDied","Data":"495097d934f0637c5c429b5cc33065b851a4cdf31cc74ed9019d3ec583ef0d65"} Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.919391 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-cw95b" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.919402 4745 scope.go:117] "RemoveContainer" containerID="716e3b93ed72d18ec3c1b8c14604171e88d17bc03cdc9ee1e0c8325bf6fbd522" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.923125 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7d87ab1d-2e84-448a-ad4e-9b672d6922a6" (UID: "7d87ab1d-2e84-448a-ad4e-9b672d6922a6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.923252 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" event={"ID":"2abd8e35-90b0-49c6-935e-2b757228070b","Type":"ContainerStarted","Data":"897fc06e54b380cf61aae7fe4f88536659f49f95c27e8cbad74e4d5b65925224"} Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.926938 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66b896bb7f-64vcw" event={"ID":"17718d89-5706-452c-a482-027b2c6f7c0d","Type":"ContainerStarted","Data":"6bc8a5904b5bb03d069ec0a58dc80f185db938ce6f84821a384b5d5b4aa771de"} Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.933536 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" event={"ID":"581b093c-c14e-499e-b5da-5591754ad44a","Type":"ContainerStarted","Data":"50b4972ec436b0a4b4806e21b13f3ed30654479365c35b8cea55080e4556c920"} Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.944112 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.944145 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.944154 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5fvc\" (UniqueName: \"kubernetes.io/projected/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-kube-api-access-h5fvc\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.944165 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.944174 4745 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.944181 4745 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d87ab1d-2e84-448a-ad4e-9b672d6922a6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.948051 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nwttm" event={"ID":"643ad5a5-986a-4003-a1ef-a0fc8a6cae10","Type":"ContainerStarted","Data":"2b09594809465432cb7afe73a35b88e8005ec15306a31a7ec1716e0da7725ab4"} Oct 10 13:33:22 crc kubenswrapper[4745]: I1010 13:33:22.953821 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.274659 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-cw95b"] Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.282027 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-cw95b"] Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.310109 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.693115 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66b896bb7f-64vcw"] Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.707341 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.729372 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.739828 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-66886d68fc-xt5ng"] Oct 10 13:33:23 crc kubenswrapper[4745]: E1010 13:33:23.740186 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d87ab1d-2e84-448a-ad4e-9b672d6922a6" containerName="init" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.740202 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d87ab1d-2e84-448a-ad4e-9b672d6922a6" containerName="init" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.740347 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d87ab1d-2e84-448a-ad4e-9b672d6922a6" containerName="init" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.741369 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.762064 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66886d68fc-xt5ng"] Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.819435 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.859547 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c688f752-0686-412c-85f8-8304cc245608-config-data\") pod \"horizon-66886d68fc-xt5ng\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.859608 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c688f752-0686-412c-85f8-8304cc245608-logs\") pod \"horizon-66886d68fc-xt5ng\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.859652 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c688f752-0686-412c-85f8-8304cc245608-scripts\") pod \"horizon-66886d68fc-xt5ng\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.859793 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vlfx\" (UniqueName: \"kubernetes.io/projected/c688f752-0686-412c-85f8-8304cc245608-kube-api-access-9vlfx\") pod \"horizon-66886d68fc-xt5ng\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.859818 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c688f752-0686-412c-85f8-8304cc245608-horizon-secret-key\") pod \"horizon-66886d68fc-xt5ng\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.956576 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5e0c23f6-8316-4e64-8d05-02d984fdb66e","Type":"ContainerStarted","Data":"d897d7a0316dc6e1c146c997cb2959428ba86488851104968388e75eb66ec055"} Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.957976 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"122e3abe-be44-46cc-826f-938bb985e0b2","Type":"ContainerStarted","Data":"797444bcc9ca1460f95c0d36282a63ece76fc9799112b28b2ee945fd50856556"} Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.959270 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-c2vrp" event={"ID":"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58","Type":"ContainerStarted","Data":"4ad2f8ae54ad3d4d5468a74615131926a7ce5eaf5a65eb17bd8419fbd905baa0"} Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.960866 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c688f752-0686-412c-85f8-8304cc245608-config-data\") pod \"horizon-66886d68fc-xt5ng\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.960903 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c688f752-0686-412c-85f8-8304cc245608-logs\") pod \"horizon-66886d68fc-xt5ng\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.960948 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c688f752-0686-412c-85f8-8304cc245608-scripts\") pod \"horizon-66886d68fc-xt5ng\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.961034 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vlfx\" (UniqueName: \"kubernetes.io/projected/c688f752-0686-412c-85f8-8304cc245608-kube-api-access-9vlfx\") pod \"horizon-66886d68fc-xt5ng\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.961065 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c688f752-0686-412c-85f8-8304cc245608-horizon-secret-key\") pod \"horizon-66886d68fc-xt5ng\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.961317 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c688f752-0686-412c-85f8-8304cc245608-logs\") pod \"horizon-66886d68fc-xt5ng\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.961745 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c688f752-0686-412c-85f8-8304cc245608-scripts\") pod \"horizon-66886d68fc-xt5ng\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.962437 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c688f752-0686-412c-85f8-8304cc245608-config-data\") pod \"horizon-66886d68fc-xt5ng\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.966765 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c688f752-0686-412c-85f8-8304cc245608-horizon-secret-key\") pod \"horizon-66886d68fc-xt5ng\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:23 crc kubenswrapper[4745]: I1010 13:33:23.979114 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vlfx\" (UniqueName: \"kubernetes.io/projected/c688f752-0686-412c-85f8-8304cc245608-kube-api-access-9vlfx\") pod \"horizon-66886d68fc-xt5ng\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:24 crc kubenswrapper[4745]: I1010 13:33:24.057744 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:24 crc kubenswrapper[4745]: I1010 13:33:24.591257 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66886d68fc-xt5ng"] Oct 10 13:33:24 crc kubenswrapper[4745]: I1010 13:33:24.761300 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d87ab1d-2e84-448a-ad4e-9b672d6922a6" path="/var/lib/kubelet/pods/7d87ab1d-2e84-448a-ad4e-9b672d6922a6/volumes" Oct 10 13:33:24 crc kubenswrapper[4745]: I1010 13:33:24.969245 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66886d68fc-xt5ng" event={"ID":"c688f752-0686-412c-85f8-8304cc245608","Type":"ContainerStarted","Data":"b38149694aab57e091d93d19e32a6fd16eafa00dd1161a676771b76b70dfb84e"} Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.319827 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-c875-account-create-mrv46"] Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.320839 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c875-account-create-mrv46" Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.323342 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.346815 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-c875-account-create-mrv46"] Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.486275 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkwpc\" (UniqueName: \"kubernetes.io/projected/a27645b6-74d3-4a15-ab07-c9767ef55413-kube-api-access-bkwpc\") pod \"cinder-c875-account-create-mrv46\" (UID: \"a27645b6-74d3-4a15-ab07-c9767ef55413\") " pod="openstack/cinder-c875-account-create-mrv46" Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.525481 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-98a4-account-create-r4qqn"] Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.527578 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-98a4-account-create-r4qqn" Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.532065 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.533016 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-98a4-account-create-r4qqn"] Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.588256 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkwpc\" (UniqueName: \"kubernetes.io/projected/a27645b6-74d3-4a15-ab07-c9767ef55413-kube-api-access-bkwpc\") pod \"cinder-c875-account-create-mrv46\" (UID: \"a27645b6-74d3-4a15-ab07-c9767ef55413\") " pod="openstack/cinder-c875-account-create-mrv46" Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.617256 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkwpc\" (UniqueName: \"kubernetes.io/projected/a27645b6-74d3-4a15-ab07-c9767ef55413-kube-api-access-bkwpc\") pod \"cinder-c875-account-create-mrv46\" (UID: \"a27645b6-74d3-4a15-ab07-c9767ef55413\") " pod="openstack/cinder-c875-account-create-mrv46" Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.688430 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c875-account-create-mrv46" Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.690028 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq5gp\" (UniqueName: \"kubernetes.io/projected/235509a2-b516-432d-a681-9db08de1e7eb-kube-api-access-cq5gp\") pod \"barbican-98a4-account-create-r4qqn\" (UID: \"235509a2-b516-432d-a681-9db08de1e7eb\") " pod="openstack/barbican-98a4-account-create-r4qqn" Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.716972 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-09e0-account-create-qsb99"] Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.718079 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-09e0-account-create-qsb99" Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.719776 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.732304 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-09e0-account-create-qsb99"] Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.791768 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq5gp\" (UniqueName: \"kubernetes.io/projected/235509a2-b516-432d-a681-9db08de1e7eb-kube-api-access-cq5gp\") pod \"barbican-98a4-account-create-r4qqn\" (UID: \"235509a2-b516-432d-a681-9db08de1e7eb\") " pod="openstack/barbican-98a4-account-create-r4qqn" Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.810613 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq5gp\" (UniqueName: \"kubernetes.io/projected/235509a2-b516-432d-a681-9db08de1e7eb-kube-api-access-cq5gp\") pod \"barbican-98a4-account-create-r4qqn\" (UID: \"235509a2-b516-432d-a681-9db08de1e7eb\") " pod="openstack/barbican-98a4-account-create-r4qqn" Oct 10 13:33:25 crc kubenswrapper[4745]: I1010 13:33:25.845210 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-98a4-account-create-r4qqn" Oct 10 13:33:26 crc kubenswrapper[4745]: I1010 13:33:25.893873 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q6kb\" (UniqueName: \"kubernetes.io/projected/335e23ac-b3dc-4bc0-b7e1-505deb67af27-kube-api-access-2q6kb\") pod \"neutron-09e0-account-create-qsb99\" (UID: \"335e23ac-b3dc-4bc0-b7e1-505deb67af27\") " pod="openstack/neutron-09e0-account-create-qsb99" Oct 10 13:33:26 crc kubenswrapper[4745]: I1010 13:33:25.995753 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q6kb\" (UniqueName: \"kubernetes.io/projected/335e23ac-b3dc-4bc0-b7e1-505deb67af27-kube-api-access-2q6kb\") pod \"neutron-09e0-account-create-qsb99\" (UID: \"335e23ac-b3dc-4bc0-b7e1-505deb67af27\") " pod="openstack/neutron-09e0-account-create-qsb99" Oct 10 13:33:26 crc kubenswrapper[4745]: I1010 13:33:26.013379 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q6kb\" (UniqueName: \"kubernetes.io/projected/335e23ac-b3dc-4bc0-b7e1-505deb67af27-kube-api-access-2q6kb\") pod \"neutron-09e0-account-create-qsb99\" (UID: \"335e23ac-b3dc-4bc0-b7e1-505deb67af27\") " pod="openstack/neutron-09e0-account-create-qsb99" Oct 10 13:33:26 crc kubenswrapper[4745]: I1010 13:33:26.213022 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-09e0-account-create-qsb99" Oct 10 13:33:26 crc kubenswrapper[4745]: I1010 13:33:26.991895 4745 generic.go:334] "Generic (PLEG): container finished" podID="2abd8e35-90b0-49c6-935e-2b757228070b" containerID="fe6923548dc550c4b9297c117d5d4c0d356b9280c25223c732f7c5d716df1791" exitCode=0 Oct 10 13:33:26 crc kubenswrapper[4745]: I1010 13:33:26.992501 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" event={"ID":"2abd8e35-90b0-49c6-935e-2b757228070b","Type":"ContainerDied","Data":"fe6923548dc550c4b9297c117d5d4c0d356b9280c25223c732f7c5d716df1791"} Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.000533 4745 generic.go:334] "Generic (PLEG): container finished" podID="581b093c-c14e-499e-b5da-5591754ad44a" containerID="6e0847159f0523b0efe0dcb201212150f1111b2701fb16e06ef3e81dc933c918" exitCode=0 Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.000598 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" event={"ID":"581b093c-c14e-499e-b5da-5591754ad44a","Type":"ContainerDied","Data":"6e0847159f0523b0efe0dcb201212150f1111b2701fb16e06ef3e81dc933c918"} Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.003365 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5e0c23f6-8316-4e64-8d05-02d984fdb66e","Type":"ContainerStarted","Data":"4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659"} Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.003394 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5e0c23f6-8316-4e64-8d05-02d984fdb66e","Type":"ContainerStarted","Data":"0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d"} Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.003545 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5e0c23f6-8316-4e64-8d05-02d984fdb66e" containerName="glance-log" containerID="cri-o://0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d" gracePeriod=30 Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.003615 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5e0c23f6-8316-4e64-8d05-02d984fdb66e" containerName="glance-httpd" containerID="cri-o://4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659" gracePeriod=30 Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.010065 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"122e3abe-be44-46cc-826f-938bb985e0b2","Type":"ContainerStarted","Data":"935f60a55b4557fc3a9aa5989cd99b0947c2142f74f53ac34a86232fc8e2344f"} Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.043036 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nwttm" event={"ID":"643ad5a5-986a-4003-a1ef-a0fc8a6cae10","Type":"ContainerStarted","Data":"463a4a20101fbec8c7faee352813af08d95ca20f65c8903271ec715a712c7656"} Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.079194 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.079173137 podStartE2EDuration="6.079173137s" podCreationTimestamp="2025-10-10 13:33:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:33:27.051403256 +0000 UTC m=+920.949060019" watchObservedRunningTime="2025-10-10 13:33:27.079173137 +0000 UTC m=+920.976829900" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.090159 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-nwttm" podStartSLOduration=6.090141753 podStartE2EDuration="6.090141753s" podCreationTimestamp="2025-10-10 13:33:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:33:27.075142305 +0000 UTC m=+920.972799068" watchObservedRunningTime="2025-10-10 13:33:27.090141753 +0000 UTC m=+920.987798506" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.105481 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-c875-account-create-mrv46"] Oct 10 13:33:27 crc kubenswrapper[4745]: W1010 13:33:27.181457 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda27645b6_74d3_4a15_ab07_c9767ef55413.slice/crio-0b5f6b951c486eb09bf2cf9cba7f0719be4c498c01c38bab6f1ae08d55e7a6ab WatchSource:0}: Error finding container 0b5f6b951c486eb09bf2cf9cba7f0719be4c498c01c38bab6f1ae08d55e7a6ab: Status 404 returned error can't find the container with id 0b5f6b951c486eb09bf2cf9cba7f0719be4c498c01c38bab6f1ae08d55e7a6ab Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.259436 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-09e0-account-create-qsb99"] Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.271348 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-98a4-account-create-r4qqn"] Oct 10 13:33:27 crc kubenswrapper[4745]: W1010 13:33:27.309979 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod335e23ac_b3dc_4bc0_b7e1_505deb67af27.slice/crio-f2029d4599016b8ecb87cd5582b05ffac1036b7c5105ab96a806a85fa32c051a WatchSource:0}: Error finding container f2029d4599016b8ecb87cd5582b05ffac1036b7c5105ab96a806a85fa32c051a: Status 404 returned error can't find the container with id f2029d4599016b8ecb87cd5582b05ffac1036b7c5105ab96a806a85fa32c051a Oct 10 13:33:27 crc kubenswrapper[4745]: W1010 13:33:27.322217 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod235509a2_b516_432d_a681_9db08de1e7eb.slice/crio-39fee32ea40d7584631052d739bab3ed07ed448905fd5515b3d23b2a606c893a WatchSource:0}: Error finding container 39fee32ea40d7584631052d739bab3ed07ed448905fd5515b3d23b2a606c893a: Status 404 returned error can't find the container with id 39fee32ea40d7584631052d739bab3ed07ed448905fd5515b3d23b2a606c893a Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.352126 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.529201 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-dns-swift-storage-0\") pod \"581b093c-c14e-499e-b5da-5591754ad44a\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.529381 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxtk4\" (UniqueName: \"kubernetes.io/projected/581b093c-c14e-499e-b5da-5591754ad44a-kube-api-access-pxtk4\") pod \"581b093c-c14e-499e-b5da-5591754ad44a\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.529413 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-ovsdbserver-nb\") pod \"581b093c-c14e-499e-b5da-5591754ad44a\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.529433 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-dns-svc\") pod \"581b093c-c14e-499e-b5da-5591754ad44a\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.529455 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-config\") pod \"581b093c-c14e-499e-b5da-5591754ad44a\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.529507 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-ovsdbserver-sb\") pod \"581b093c-c14e-499e-b5da-5591754ad44a\" (UID: \"581b093c-c14e-499e-b5da-5591754ad44a\") " Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.540334 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/581b093c-c14e-499e-b5da-5591754ad44a-kube-api-access-pxtk4" (OuterVolumeSpecName: "kube-api-access-pxtk4") pod "581b093c-c14e-499e-b5da-5591754ad44a" (UID: "581b093c-c14e-499e-b5da-5591754ad44a"). InnerVolumeSpecName "kube-api-access-pxtk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.577612 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "581b093c-c14e-499e-b5da-5591754ad44a" (UID: "581b093c-c14e-499e-b5da-5591754ad44a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.588460 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "581b093c-c14e-499e-b5da-5591754ad44a" (UID: "581b093c-c14e-499e-b5da-5591754ad44a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.589171 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "581b093c-c14e-499e-b5da-5591754ad44a" (UID: "581b093c-c14e-499e-b5da-5591754ad44a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.595170 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-config" (OuterVolumeSpecName: "config") pod "581b093c-c14e-499e-b5da-5591754ad44a" (UID: "581b093c-c14e-499e-b5da-5591754ad44a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.595881 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "581b093c-c14e-499e-b5da-5591754ad44a" (UID: "581b093c-c14e-499e-b5da-5591754ad44a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.631454 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxtk4\" (UniqueName: \"kubernetes.io/projected/581b093c-c14e-499e-b5da-5591754ad44a-kube-api-access-pxtk4\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.631485 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.631494 4745 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.631503 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.631510 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.631519 4745 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/581b093c-c14e-499e-b5da-5591754ad44a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.804142 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.936266 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-scripts\") pod \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.936347 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.936422 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-combined-ca-bundle\") pod \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.936478 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9j4bn\" (UniqueName: \"kubernetes.io/projected/5e0c23f6-8316-4e64-8d05-02d984fdb66e-kube-api-access-9j4bn\") pod \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.937034 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e0c23f6-8316-4e64-8d05-02d984fdb66e-logs\") pod \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.937124 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-config-data\") pod \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.937157 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5e0c23f6-8316-4e64-8d05-02d984fdb66e-httpd-run\") pod \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\" (UID: \"5e0c23f6-8316-4e64-8d05-02d984fdb66e\") " Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.937893 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e0c23f6-8316-4e64-8d05-02d984fdb66e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5e0c23f6-8316-4e64-8d05-02d984fdb66e" (UID: "5e0c23f6-8316-4e64-8d05-02d984fdb66e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.938186 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e0c23f6-8316-4e64-8d05-02d984fdb66e-logs" (OuterVolumeSpecName: "logs") pod "5e0c23f6-8316-4e64-8d05-02d984fdb66e" (UID: "5e0c23f6-8316-4e64-8d05-02d984fdb66e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.948397 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e0c23f6-8316-4e64-8d05-02d984fdb66e-kube-api-access-9j4bn" (OuterVolumeSpecName: "kube-api-access-9j4bn") pod "5e0c23f6-8316-4e64-8d05-02d984fdb66e" (UID: "5e0c23f6-8316-4e64-8d05-02d984fdb66e"). InnerVolumeSpecName "kube-api-access-9j4bn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.949854 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-scripts" (OuterVolumeSpecName: "scripts") pod "5e0c23f6-8316-4e64-8d05-02d984fdb66e" (UID: "5e0c23f6-8316-4e64-8d05-02d984fdb66e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.949925 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "5e0c23f6-8316-4e64-8d05-02d984fdb66e" (UID: "5e0c23f6-8316-4e64-8d05-02d984fdb66e"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.964607 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e0c23f6-8316-4e64-8d05-02d984fdb66e" (UID: "5e0c23f6-8316-4e64-8d05-02d984fdb66e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:27 crc kubenswrapper[4745]: I1010 13:33:27.990229 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-config-data" (OuterVolumeSpecName: "config-data") pod "5e0c23f6-8316-4e64-8d05-02d984fdb66e" (UID: "5e0c23f6-8316-4e64-8d05-02d984fdb66e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.038980 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.039007 4745 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5e0c23f6-8316-4e64-8d05-02d984fdb66e-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.039017 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.039040 4745 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.039049 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e0c23f6-8316-4e64-8d05-02d984fdb66e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.039061 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9j4bn\" (UniqueName: \"kubernetes.io/projected/5e0c23f6-8316-4e64-8d05-02d984fdb66e-kube-api-access-9j4bn\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.039071 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e0c23f6-8316-4e64-8d05-02d984fdb66e-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.061215 4745 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.067915 4745 generic.go:334] "Generic (PLEG): container finished" podID="a27645b6-74d3-4a15-ab07-c9767ef55413" containerID="9497ef0ece50c843aef82bd37fb30c7a6e552c24bf36a6a35253efce4286198e" exitCode=0 Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.068043 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c875-account-create-mrv46" event={"ID":"a27645b6-74d3-4a15-ab07-c9767ef55413","Type":"ContainerDied","Data":"9497ef0ece50c843aef82bd37fb30c7a6e552c24bf36a6a35253efce4286198e"} Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.068071 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c875-account-create-mrv46" event={"ID":"a27645b6-74d3-4a15-ab07-c9767ef55413","Type":"ContainerStarted","Data":"0b5f6b951c486eb09bf2cf9cba7f0719be4c498c01c38bab6f1ae08d55e7a6ab"} Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.070962 4745 generic.go:334] "Generic (PLEG): container finished" podID="335e23ac-b3dc-4bc0-b7e1-505deb67af27" containerID="8f2c29aec06f8555ebac52308b3893ce31a90233c10d431b1255d3755d2ea170" exitCode=0 Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.070999 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-09e0-account-create-qsb99" event={"ID":"335e23ac-b3dc-4bc0-b7e1-505deb67af27","Type":"ContainerDied","Data":"8f2c29aec06f8555ebac52308b3893ce31a90233c10d431b1255d3755d2ea170"} Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.071012 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-09e0-account-create-qsb99" event={"ID":"335e23ac-b3dc-4bc0-b7e1-505deb67af27","Type":"ContainerStarted","Data":"f2029d4599016b8ecb87cd5582b05ffac1036b7c5105ab96a806a85fa32c051a"} Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.073297 4745 generic.go:334] "Generic (PLEG): container finished" podID="235509a2-b516-432d-a681-9db08de1e7eb" containerID="29be1db3925cfbfc891675f55b1fcbd20948a76fa309b906903dbecfe3f9ba1a" exitCode=0 Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.073337 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-98a4-account-create-r4qqn" event={"ID":"235509a2-b516-432d-a681-9db08de1e7eb","Type":"ContainerDied","Data":"29be1db3925cfbfc891675f55b1fcbd20948a76fa309b906903dbecfe3f9ba1a"} Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.073353 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-98a4-account-create-r4qqn" event={"ID":"235509a2-b516-432d-a681-9db08de1e7eb","Type":"ContainerStarted","Data":"39fee32ea40d7584631052d739bab3ed07ed448905fd5515b3d23b2a606c893a"} Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.078046 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" event={"ID":"2abd8e35-90b0-49c6-935e-2b757228070b","Type":"ContainerStarted","Data":"d4cc78df2a3fa8942a88a570f62105212fdec51468aac828e752d0db38ffdc8d"} Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.078909 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.081147 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" event={"ID":"581b093c-c14e-499e-b5da-5591754ad44a","Type":"ContainerDied","Data":"50b4972ec436b0a4b4806e21b13f3ed30654479365c35b8cea55080e4556c920"} Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.081184 4745 scope.go:117] "RemoveContainer" containerID="6e0847159f0523b0efe0dcb201212150f1111b2701fb16e06ef3e81dc933c918" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.081293 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.092536 4745 generic.go:334] "Generic (PLEG): container finished" podID="5e0c23f6-8316-4e64-8d05-02d984fdb66e" containerID="4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659" exitCode=143 Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.092562 4745 generic.go:334] "Generic (PLEG): container finished" podID="5e0c23f6-8316-4e64-8d05-02d984fdb66e" containerID="0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d" exitCode=143 Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.092619 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5e0c23f6-8316-4e64-8d05-02d984fdb66e","Type":"ContainerDied","Data":"4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659"} Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.092647 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5e0c23f6-8316-4e64-8d05-02d984fdb66e","Type":"ContainerDied","Data":"0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d"} Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.092660 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5e0c23f6-8316-4e64-8d05-02d984fdb66e","Type":"ContainerDied","Data":"d897d7a0316dc6e1c146c997cb2959428ba86488851104968388e75eb66ec055"} Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.092701 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.102301 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"122e3abe-be44-46cc-826f-938bb985e0b2","Type":"ContainerStarted","Data":"da871c371d2c53e62ff85cc80535422abcceefce9f165cb531db25433f469db4"} Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.102434 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="122e3abe-be44-46cc-826f-938bb985e0b2" containerName="glance-log" containerID="cri-o://935f60a55b4557fc3a9aa5989cd99b0947c2142f74f53ac34a86232fc8e2344f" gracePeriod=30 Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.102526 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="122e3abe-be44-46cc-826f-938bb985e0b2" containerName="glance-httpd" containerID="cri-o://da871c371d2c53e62ff85cc80535422abcceefce9f165cb531db25433f469db4" gracePeriod=30 Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.110298 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" podStartSLOduration=7.11028188 podStartE2EDuration="7.11028188s" podCreationTimestamp="2025-10-10 13:33:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:33:28.100856152 +0000 UTC m=+921.998512915" watchObservedRunningTime="2025-10-10 13:33:28.11028188 +0000 UTC m=+922.007938643" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.122015 4745 scope.go:117] "RemoveContainer" containerID="4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.142706 4745 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.190913 4745 scope.go:117] "RemoveContainer" containerID="0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.195854 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8"] Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.209443 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-fn2x8"] Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.218915 4745 scope.go:117] "RemoveContainer" containerID="4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659" Oct 10 13:33:28 crc kubenswrapper[4745]: E1010 13:33:28.219431 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659\": container with ID starting with 4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659 not found: ID does not exist" containerID="4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.219490 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659"} err="failed to get container status \"4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659\": rpc error: code = NotFound desc = could not find container \"4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659\": container with ID starting with 4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659 not found: ID does not exist" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.219538 4745 scope.go:117] "RemoveContainer" containerID="0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d" Oct 10 13:33:28 crc kubenswrapper[4745]: E1010 13:33:28.220050 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d\": container with ID starting with 0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d not found: ID does not exist" containerID="0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.220115 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d"} err="failed to get container status \"0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d\": rpc error: code = NotFound desc = could not find container \"0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d\": container with ID starting with 0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d not found: ID does not exist" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.220138 4745 scope.go:117] "RemoveContainer" containerID="4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.220403 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659"} err="failed to get container status \"4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659\": rpc error: code = NotFound desc = could not find container \"4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659\": container with ID starting with 4968b8c74dda05786d753d9517a00f55e91b77196779cc8498997e65b97b9659 not found: ID does not exist" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.220431 4745 scope.go:117] "RemoveContainer" containerID="0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.220784 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d"} err="failed to get container status \"0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d\": rpc error: code = NotFound desc = could not find container \"0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d\": container with ID starting with 0070b33787471f9951774600d674d40f8f6222910e6e68c5bbaf5eb7288aa66d not found: ID does not exist" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.223064 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.2230411740000005 podStartE2EDuration="7.223041174s" podCreationTimestamp="2025-10-10 13:33:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:33:28.211183225 +0000 UTC m=+922.108839988" watchObservedRunningTime="2025-10-10 13:33:28.223041174 +0000 UTC m=+922.120697937" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.256854 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.337015 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.350374 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:33:28 crc kubenswrapper[4745]: E1010 13:33:28.350774 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="581b093c-c14e-499e-b5da-5591754ad44a" containerName="init" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.350788 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="581b093c-c14e-499e-b5da-5591754ad44a" containerName="init" Oct 10 13:33:28 crc kubenswrapper[4745]: E1010 13:33:28.350807 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e0c23f6-8316-4e64-8d05-02d984fdb66e" containerName="glance-httpd" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.350813 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e0c23f6-8316-4e64-8d05-02d984fdb66e" containerName="glance-httpd" Oct 10 13:33:28 crc kubenswrapper[4745]: E1010 13:33:28.350835 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e0c23f6-8316-4e64-8d05-02d984fdb66e" containerName="glance-log" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.350841 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e0c23f6-8316-4e64-8d05-02d984fdb66e" containerName="glance-log" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.351020 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="581b093c-c14e-499e-b5da-5591754ad44a" containerName="init" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.351039 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e0c23f6-8316-4e64-8d05-02d984fdb66e" containerName="glance-log" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.351052 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e0c23f6-8316-4e64-8d05-02d984fdb66e" containerName="glance-httpd" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.351938 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.353768 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.360026 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.453540 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tlc9\" (UniqueName: \"kubernetes.io/projected/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-kube-api-access-8tlc9\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.453626 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-logs\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.453654 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.453680 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.453720 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.453777 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-config-data\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.453801 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-scripts\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.555044 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tlc9\" (UniqueName: \"kubernetes.io/projected/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-kube-api-access-8tlc9\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.555122 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-logs\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.555148 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.555161 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.555199 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.555218 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-config-data\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.555243 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-scripts\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.555901 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.556098 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.556234 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-logs\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.560964 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.564249 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-config-data\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.569365 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-scripts\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.573444 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tlc9\" (UniqueName: \"kubernetes.io/projected/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-kube-api-access-8tlc9\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.592983 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.679857 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.760583 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="581b093c-c14e-499e-b5da-5591754ad44a" path="/var/lib/kubelet/pods/581b093c-c14e-499e-b5da-5591754ad44a/volumes" Oct 10 13:33:28 crc kubenswrapper[4745]: I1010 13:33:28.761245 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e0c23f6-8316-4e64-8d05-02d984fdb66e" path="/var/lib/kubelet/pods/5e0c23f6-8316-4e64-8d05-02d984fdb66e/volumes" Oct 10 13:33:29 crc kubenswrapper[4745]: I1010 13:33:29.123720 4745 generic.go:334] "Generic (PLEG): container finished" podID="122e3abe-be44-46cc-826f-938bb985e0b2" containerID="da871c371d2c53e62ff85cc80535422abcceefce9f165cb531db25433f469db4" exitCode=0 Oct 10 13:33:29 crc kubenswrapper[4745]: I1010 13:33:29.124038 4745 generic.go:334] "Generic (PLEG): container finished" podID="122e3abe-be44-46cc-826f-938bb985e0b2" containerID="935f60a55b4557fc3a9aa5989cd99b0947c2142f74f53ac34a86232fc8e2344f" exitCode=143 Oct 10 13:33:29 crc kubenswrapper[4745]: I1010 13:33:29.123756 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"122e3abe-be44-46cc-826f-938bb985e0b2","Type":"ContainerDied","Data":"da871c371d2c53e62ff85cc80535422abcceefce9f165cb531db25433f469db4"} Oct 10 13:33:29 crc kubenswrapper[4745]: I1010 13:33:29.124169 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"122e3abe-be44-46cc-826f-938bb985e0b2","Type":"ContainerDied","Data":"935f60a55b4557fc3a9aa5989cd99b0947c2142f74f53ac34a86232fc8e2344f"} Oct 10 13:33:30 crc kubenswrapper[4745]: I1010 13:33:30.133794 4745 generic.go:334] "Generic (PLEG): container finished" podID="643ad5a5-986a-4003-a1ef-a0fc8a6cae10" containerID="463a4a20101fbec8c7faee352813af08d95ca20f65c8903271ec715a712c7656" exitCode=0 Oct 10 13:33:30 crc kubenswrapper[4745]: I1010 13:33:30.133851 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nwttm" event={"ID":"643ad5a5-986a-4003-a1ef-a0fc8a6cae10","Type":"ContainerDied","Data":"463a4a20101fbec8c7faee352813af08d95ca20f65c8903271ec715a712c7656"} Oct 10 13:33:30 crc kubenswrapper[4745]: I1010 13:33:30.530522 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-09e0-account-create-qsb99" Oct 10 13:33:30 crc kubenswrapper[4745]: I1010 13:33:30.702292 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2q6kb\" (UniqueName: \"kubernetes.io/projected/335e23ac-b3dc-4bc0-b7e1-505deb67af27-kube-api-access-2q6kb\") pod \"335e23ac-b3dc-4bc0-b7e1-505deb67af27\" (UID: \"335e23ac-b3dc-4bc0-b7e1-505deb67af27\") " Oct 10 13:33:30 crc kubenswrapper[4745]: I1010 13:33:30.712958 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/335e23ac-b3dc-4bc0-b7e1-505deb67af27-kube-api-access-2q6kb" (OuterVolumeSpecName: "kube-api-access-2q6kb") pod "335e23ac-b3dc-4bc0-b7e1-505deb67af27" (UID: "335e23ac-b3dc-4bc0-b7e1-505deb67af27"). InnerVolumeSpecName "kube-api-access-2q6kb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:33:30 crc kubenswrapper[4745]: I1010 13:33:30.803750 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2q6kb\" (UniqueName: \"kubernetes.io/projected/335e23ac-b3dc-4bc0-b7e1-505deb67af27-kube-api-access-2q6kb\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:31 crc kubenswrapper[4745]: I1010 13:33:31.160488 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-09e0-account-create-qsb99" Oct 10 13:33:31 crc kubenswrapper[4745]: I1010 13:33:31.160914 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-09e0-account-create-qsb99" event={"ID":"335e23ac-b3dc-4bc0-b7e1-505deb67af27","Type":"ContainerDied","Data":"f2029d4599016b8ecb87cd5582b05ffac1036b7c5105ab96a806a85fa32c051a"} Oct 10 13:33:31 crc kubenswrapper[4745]: I1010 13:33:31.160974 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2029d4599016b8ecb87cd5582b05ffac1036b7c5105ab96a806a85fa32c051a" Oct 10 13:33:32 crc kubenswrapper[4745]: I1010 13:33:32.022270 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:33:32 crc kubenswrapper[4745]: I1010 13:33:32.138947 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-nv8xl"] Oct 10 13:33:32 crc kubenswrapper[4745]: I1010 13:33:32.139213 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" podUID="b683a409-b7d3-404f-ac5a-43be34e071c6" containerName="dnsmasq-dns" containerID="cri-o://94760b0459376bca580d88cb72faa38ce3eb1269d08ffebd83ce22fcc54fb3ff" gracePeriod=10 Oct 10 13:33:32 crc kubenswrapper[4745]: I1010 13:33:32.153592 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.032692 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6fcdc9c87c-9r2nf"] Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.076285 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-65c7d6f5bb-9g4xm"] Oct 10 13:33:33 crc kubenswrapper[4745]: E1010 13:33:33.076777 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="335e23ac-b3dc-4bc0-b7e1-505deb67af27" containerName="mariadb-account-create" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.076796 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="335e23ac-b3dc-4bc0-b7e1-505deb67af27" containerName="mariadb-account-create" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.076963 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="335e23ac-b3dc-4bc0-b7e1-505deb67af27" containerName="mariadb-account-create" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.079054 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.085543 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.092965 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-65c7d6f5bb-9g4xm"] Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.214571 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66886d68fc-xt5ng"] Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.217247 4745 generic.go:334] "Generic (PLEG): container finished" podID="b683a409-b7d3-404f-ac5a-43be34e071c6" containerID="94760b0459376bca580d88cb72faa38ce3eb1269d08ffebd83ce22fcc54fb3ff" exitCode=0 Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.217301 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" event={"ID":"b683a409-b7d3-404f-ac5a-43be34e071c6","Type":"ContainerDied","Data":"94760b0459376bca580d88cb72faa38ce3eb1269d08ffebd83ce22fcc54fb3ff"} Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.251587 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-8574589b48-pc9wr"] Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.253763 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.254973 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftrmk\" (UniqueName: \"kubernetes.io/projected/604fc859-90dd-4c8b-a216-6009b3dbcbf2-kube-api-access-ftrmk\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.255032 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/604fc859-90dd-4c8b-a216-6009b3dbcbf2-logs\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.255060 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-horizon-secret-key\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.255092 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-combined-ca-bundle\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.255122 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-horizon-tls-certs\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.255168 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/604fc859-90dd-4c8b-a216-6009b3dbcbf2-scripts\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.255192 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/604fc859-90dd-4c8b-a216-6009b3dbcbf2-config-data\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.269750 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8574589b48-pc9wr"] Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.357514 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/604fc859-90dd-4c8b-a216-6009b3dbcbf2-scripts\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.357557 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb40990e-1db6-4eab-9365-06678441a9ba-horizon-tls-certs\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.357659 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/604fc859-90dd-4c8b-a216-6009b3dbcbf2-config-data\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.357686 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftrmk\" (UniqueName: \"kubernetes.io/projected/604fc859-90dd-4c8b-a216-6009b3dbcbf2-kube-api-access-ftrmk\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.357719 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb40990e-1db6-4eab-9365-06678441a9ba-logs\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.357756 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkbsg\" (UniqueName: \"kubernetes.io/projected/fb40990e-1db6-4eab-9365-06678441a9ba-kube-api-access-lkbsg\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.357772 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb40990e-1db6-4eab-9365-06678441a9ba-combined-ca-bundle\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.357791 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb40990e-1db6-4eab-9365-06678441a9ba-horizon-secret-key\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.357814 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/604fc859-90dd-4c8b-a216-6009b3dbcbf2-logs\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.357837 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-horizon-secret-key\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.357853 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb40990e-1db6-4eab-9365-06678441a9ba-config-data\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.357896 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb40990e-1db6-4eab-9365-06678441a9ba-scripts\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.357926 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-combined-ca-bundle\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.357973 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-horizon-tls-certs\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.359316 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/604fc859-90dd-4c8b-a216-6009b3dbcbf2-logs\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.359439 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/604fc859-90dd-4c8b-a216-6009b3dbcbf2-scripts\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.360418 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/604fc859-90dd-4c8b-a216-6009b3dbcbf2-config-data\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.366497 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-horizon-tls-certs\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.367117 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-combined-ca-bundle\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.377426 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-horizon-secret-key\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.380145 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftrmk\" (UniqueName: \"kubernetes.io/projected/604fc859-90dd-4c8b-a216-6009b3dbcbf2-kube-api-access-ftrmk\") pod \"horizon-65c7d6f5bb-9g4xm\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.411159 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.459173 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb40990e-1db6-4eab-9365-06678441a9ba-horizon-tls-certs\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.459251 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb40990e-1db6-4eab-9365-06678441a9ba-logs\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.459293 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkbsg\" (UniqueName: \"kubernetes.io/projected/fb40990e-1db6-4eab-9365-06678441a9ba-kube-api-access-lkbsg\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.459316 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb40990e-1db6-4eab-9365-06678441a9ba-combined-ca-bundle\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.459341 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb40990e-1db6-4eab-9365-06678441a9ba-horizon-secret-key\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.459390 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb40990e-1db6-4eab-9365-06678441a9ba-config-data\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.459418 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb40990e-1db6-4eab-9365-06678441a9ba-scripts\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.460342 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb40990e-1db6-4eab-9365-06678441a9ba-scripts\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.462039 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb40990e-1db6-4eab-9365-06678441a9ba-logs\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.463530 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb40990e-1db6-4eab-9365-06678441a9ba-config-data\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.466677 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb40990e-1db6-4eab-9365-06678441a9ba-horizon-tls-certs\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.467184 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb40990e-1db6-4eab-9365-06678441a9ba-combined-ca-bundle\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.478794 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb40990e-1db6-4eab-9365-06678441a9ba-horizon-secret-key\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.480981 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkbsg\" (UniqueName: \"kubernetes.io/projected/fb40990e-1db6-4eab-9365-06678441a9ba-kube-api-access-lkbsg\") pod \"horizon-8574589b48-pc9wr\" (UID: \"fb40990e-1db6-4eab-9365-06678441a9ba\") " pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:33 crc kubenswrapper[4745]: I1010 13:33:33.577136 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:35 crc kubenswrapper[4745]: I1010 13:33:35.906175 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-krvls"] Oct 10 13:33:35 crc kubenswrapper[4745]: I1010 13:33:35.908181 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-krvls" Oct 10 13:33:35 crc kubenswrapper[4745]: I1010 13:33:35.910416 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 10 13:33:35 crc kubenswrapper[4745]: I1010 13:33:35.910679 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 10 13:33:35 crc kubenswrapper[4745]: I1010 13:33:35.910768 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-wdk48" Oct 10 13:33:35 crc kubenswrapper[4745]: I1010 13:33:35.924075 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-krvls"] Oct 10 13:33:36 crc kubenswrapper[4745]: I1010 13:33:36.003343 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wct7m\" (UniqueName: \"kubernetes.io/projected/45009f2e-d0d5-4a20-b62c-d5d0534e1551-kube-api-access-wct7m\") pod \"neutron-db-sync-krvls\" (UID: \"45009f2e-d0d5-4a20-b62c-d5d0534e1551\") " pod="openstack/neutron-db-sync-krvls" Oct 10 13:33:36 crc kubenswrapper[4745]: I1010 13:33:36.003415 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45009f2e-d0d5-4a20-b62c-d5d0534e1551-combined-ca-bundle\") pod \"neutron-db-sync-krvls\" (UID: \"45009f2e-d0d5-4a20-b62c-d5d0534e1551\") " pod="openstack/neutron-db-sync-krvls" Oct 10 13:33:36 crc kubenswrapper[4745]: I1010 13:33:36.003438 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/45009f2e-d0d5-4a20-b62c-d5d0534e1551-config\") pod \"neutron-db-sync-krvls\" (UID: \"45009f2e-d0d5-4a20-b62c-d5d0534e1551\") " pod="openstack/neutron-db-sync-krvls" Oct 10 13:33:36 crc kubenswrapper[4745]: I1010 13:33:36.105292 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45009f2e-d0d5-4a20-b62c-d5d0534e1551-combined-ca-bundle\") pod \"neutron-db-sync-krvls\" (UID: \"45009f2e-d0d5-4a20-b62c-d5d0534e1551\") " pod="openstack/neutron-db-sync-krvls" Oct 10 13:33:36 crc kubenswrapper[4745]: I1010 13:33:36.105363 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/45009f2e-d0d5-4a20-b62c-d5d0534e1551-config\") pod \"neutron-db-sync-krvls\" (UID: \"45009f2e-d0d5-4a20-b62c-d5d0534e1551\") " pod="openstack/neutron-db-sync-krvls" Oct 10 13:33:36 crc kubenswrapper[4745]: I1010 13:33:36.105565 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wct7m\" (UniqueName: \"kubernetes.io/projected/45009f2e-d0d5-4a20-b62c-d5d0534e1551-kube-api-access-wct7m\") pod \"neutron-db-sync-krvls\" (UID: \"45009f2e-d0d5-4a20-b62c-d5d0534e1551\") " pod="openstack/neutron-db-sync-krvls" Oct 10 13:33:36 crc kubenswrapper[4745]: I1010 13:33:36.111385 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/45009f2e-d0d5-4a20-b62c-d5d0534e1551-config\") pod \"neutron-db-sync-krvls\" (UID: \"45009f2e-d0d5-4a20-b62c-d5d0534e1551\") " pod="openstack/neutron-db-sync-krvls" Oct 10 13:33:36 crc kubenswrapper[4745]: I1010 13:33:36.111499 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45009f2e-d0d5-4a20-b62c-d5d0534e1551-combined-ca-bundle\") pod \"neutron-db-sync-krvls\" (UID: \"45009f2e-d0d5-4a20-b62c-d5d0534e1551\") " pod="openstack/neutron-db-sync-krvls" Oct 10 13:33:36 crc kubenswrapper[4745]: I1010 13:33:36.128022 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wct7m\" (UniqueName: \"kubernetes.io/projected/45009f2e-d0d5-4a20-b62c-d5d0534e1551-kube-api-access-wct7m\") pod \"neutron-db-sync-krvls\" (UID: \"45009f2e-d0d5-4a20-b62c-d5d0534e1551\") " pod="openstack/neutron-db-sync-krvls" Oct 10 13:33:36 crc kubenswrapper[4745]: I1010 13:33:36.194914 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" podUID="b683a409-b7d3-404f-ac5a-43be34e071c6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.124:5353: connect: connection refused" Oct 10 13:33:36 crc kubenswrapper[4745]: I1010 13:33:36.243799 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-krvls" Oct 10 13:33:39 crc kubenswrapper[4745]: E1010 13:33:39.574685 4745 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Oct 10 13:33:39 crc kubenswrapper[4745]: E1010 13:33:39.575262 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5w9cx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-c2vrp_openstack(60ccdbf7-fc2d-4292-8ff6-d388ccda2b58): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 13:33:39 crc kubenswrapper[4745]: E1010 13:33:39.576546 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-c2vrp" podUID="60ccdbf7-fc2d-4292-8ff6-d388ccda2b58" Oct 10 13:33:39 crc kubenswrapper[4745]: E1010 13:33:39.600327 4745 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 10 13:33:39 crc kubenswrapper[4745]: E1010 13:33:39.600526 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ncbh584h54bhd4h584hcbh59dh599h6ch5b4h5ch5cfh9chch5ddh85hb5h694h54h65fh54dh7fh587h645h5d5h5fh667h566h697h5dch58dh59fq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mj2lz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-66b896bb7f-64vcw_openstack(17718d89-5706-452c-a482-027b2c6f7c0d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 13:33:39 crc kubenswrapper[4745]: E1010 13:33:39.631698 4745 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 10 13:33:39 crc kubenswrapper[4745]: E1010 13:33:39.631911 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n599h8dhbfh57bh5cch5c8h67chf4h5bdh568h697hd7hf5h64bh64fh588hfbh68chdfh666h5h79h565h577h544h654hd5h5cchbhdh68dh67fq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9vlfx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-66886d68fc-xt5ng_openstack(c688f752-0686-412c-85f8-8304cc245608): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.697127 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-98a4-account-create-r4qqn" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.706055 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.867350 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c875-account-create-mrv46" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.872408 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.879249 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-combined-ca-bundle\") pod \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.879285 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-credential-keys\") pod \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.879307 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q9rg\" (UniqueName: \"kubernetes.io/projected/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-kube-api-access-6q9rg\") pod \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.881921 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgkcx\" (UniqueName: \"kubernetes.io/projected/122e3abe-be44-46cc-826f-938bb985e0b2-kube-api-access-mgkcx\") pod \"122e3abe-be44-46cc-826f-938bb985e0b2\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.881954 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkwpc\" (UniqueName: \"kubernetes.io/projected/a27645b6-74d3-4a15-ab07-c9767ef55413-kube-api-access-bkwpc\") pod \"a27645b6-74d3-4a15-ab07-c9767ef55413\" (UID: \"a27645b6-74d3-4a15-ab07-c9767ef55413\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.881973 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-scripts\") pod \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.882032 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-config-data\") pod \"122e3abe-be44-46cc-826f-938bb985e0b2\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.882094 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-fernet-keys\") pod \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.882118 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-combined-ca-bundle\") pod \"122e3abe-be44-46cc-826f-938bb985e0b2\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.882143 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-config-data\") pod \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\" (UID: \"643ad5a5-986a-4003-a1ef-a0fc8a6cae10\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.882184 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"122e3abe-be44-46cc-826f-938bb985e0b2\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.882212 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cq5gp\" (UniqueName: \"kubernetes.io/projected/235509a2-b516-432d-a681-9db08de1e7eb-kube-api-access-cq5gp\") pod \"235509a2-b516-432d-a681-9db08de1e7eb\" (UID: \"235509a2-b516-432d-a681-9db08de1e7eb\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.882230 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122e3abe-be44-46cc-826f-938bb985e0b2-logs\") pod \"122e3abe-be44-46cc-826f-938bb985e0b2\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.882254 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-scripts\") pod \"122e3abe-be44-46cc-826f-938bb985e0b2\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.882280 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/122e3abe-be44-46cc-826f-938bb985e0b2-httpd-run\") pod \"122e3abe-be44-46cc-826f-938bb985e0b2\" (UID: \"122e3abe-be44-46cc-826f-938bb985e0b2\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.883500 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/122e3abe-be44-46cc-826f-938bb985e0b2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "122e3abe-be44-46cc-826f-938bb985e0b2" (UID: "122e3abe-be44-46cc-826f-938bb985e0b2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.890714 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "643ad5a5-986a-4003-a1ef-a0fc8a6cae10" (UID: "643ad5a5-986a-4003-a1ef-a0fc8a6cae10"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.895355 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-kube-api-access-6q9rg" (OuterVolumeSpecName: "kube-api-access-6q9rg") pod "643ad5a5-986a-4003-a1ef-a0fc8a6cae10" (UID: "643ad5a5-986a-4003-a1ef-a0fc8a6cae10"). InnerVolumeSpecName "kube-api-access-6q9rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.902642 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/235509a2-b516-432d-a681-9db08de1e7eb-kube-api-access-cq5gp" (OuterVolumeSpecName: "kube-api-access-cq5gp") pod "235509a2-b516-432d-a681-9db08de1e7eb" (UID: "235509a2-b516-432d-a681-9db08de1e7eb"). InnerVolumeSpecName "kube-api-access-cq5gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.906154 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "122e3abe-be44-46cc-826f-938bb985e0b2" (UID: "122e3abe-be44-46cc-826f-938bb985e0b2"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.908948 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a27645b6-74d3-4a15-ab07-c9767ef55413-kube-api-access-bkwpc" (OuterVolumeSpecName: "kube-api-access-bkwpc") pod "a27645b6-74d3-4a15-ab07-c9767ef55413" (UID: "a27645b6-74d3-4a15-ab07-c9767ef55413"). InnerVolumeSpecName "kube-api-access-bkwpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.911104 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-scripts" (OuterVolumeSpecName: "scripts") pod "122e3abe-be44-46cc-826f-938bb985e0b2" (UID: "122e3abe-be44-46cc-826f-938bb985e0b2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.911495 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "643ad5a5-986a-4003-a1ef-a0fc8a6cae10" (UID: "643ad5a5-986a-4003-a1ef-a0fc8a6cae10"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.911651 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/122e3abe-be44-46cc-826f-938bb985e0b2-logs" (OuterVolumeSpecName: "logs") pod "122e3abe-be44-46cc-826f-938bb985e0b2" (UID: "122e3abe-be44-46cc-826f-938bb985e0b2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.918112 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/122e3abe-be44-46cc-826f-938bb985e0b2-kube-api-access-mgkcx" (OuterVolumeSpecName: "kube-api-access-mgkcx") pod "122e3abe-be44-46cc-826f-938bb985e0b2" (UID: "122e3abe-be44-46cc-826f-938bb985e0b2"). InnerVolumeSpecName "kube-api-access-mgkcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.918205 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-scripts" (OuterVolumeSpecName: "scripts") pod "643ad5a5-986a-4003-a1ef-a0fc8a6cae10" (UID: "643ad5a5-986a-4003-a1ef-a0fc8a6cae10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.964209 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-config-data" (OuterVolumeSpecName: "config-data") pod "643ad5a5-986a-4003-a1ef-a0fc8a6cae10" (UID: "643ad5a5-986a-4003-a1ef-a0fc8a6cae10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.977475 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.983289 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-dns-swift-storage-0\") pod \"b683a409-b7d3-404f-ac5a-43be34e071c6\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.983374 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbmh5\" (UniqueName: \"kubernetes.io/projected/b683a409-b7d3-404f-ac5a-43be34e071c6-kube-api-access-mbmh5\") pod \"b683a409-b7d3-404f-ac5a-43be34e071c6\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.983416 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-dns-svc\") pod \"b683a409-b7d3-404f-ac5a-43be34e071c6\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.983484 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-ovsdbserver-nb\") pod \"b683a409-b7d3-404f-ac5a-43be34e071c6\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.983623 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-ovsdbserver-sb\") pod \"b683a409-b7d3-404f-ac5a-43be34e071c6\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.983670 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-config\") pod \"b683a409-b7d3-404f-ac5a-43be34e071c6\" (UID: \"b683a409-b7d3-404f-ac5a-43be34e071c6\") " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.984078 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.984110 4745 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.984121 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cq5gp\" (UniqueName: \"kubernetes.io/projected/235509a2-b516-432d-a681-9db08de1e7eb-kube-api-access-cq5gp\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.984131 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122e3abe-be44-46cc-826f-938bb985e0b2-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.984139 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.984147 4745 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/122e3abe-be44-46cc-826f-938bb985e0b2-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.984155 4745 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.984165 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6q9rg\" (UniqueName: \"kubernetes.io/projected/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-kube-api-access-6q9rg\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.984174 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgkcx\" (UniqueName: \"kubernetes.io/projected/122e3abe-be44-46cc-826f-938bb985e0b2-kube-api-access-mgkcx\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.984184 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkwpc\" (UniqueName: \"kubernetes.io/projected/a27645b6-74d3-4a15-ab07-c9767ef55413-kube-api-access-bkwpc\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.984194 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:39 crc kubenswrapper[4745]: I1010 13:33:39.984201 4745 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.012776 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b683a409-b7d3-404f-ac5a-43be34e071c6-kube-api-access-mbmh5" (OuterVolumeSpecName: "kube-api-access-mbmh5") pod "b683a409-b7d3-404f-ac5a-43be34e071c6" (UID: "b683a409-b7d3-404f-ac5a-43be34e071c6"). InnerVolumeSpecName "kube-api-access-mbmh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.060720 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "643ad5a5-986a-4003-a1ef-a0fc8a6cae10" (UID: "643ad5a5-986a-4003-a1ef-a0fc8a6cae10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:40 crc kubenswrapper[4745]: E1010 13:33:40.064921 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/horizon-66886d68fc-xt5ng" podUID="c688f752-0686-412c-85f8-8304cc245608" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.086810 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbmh5\" (UniqueName: \"kubernetes.io/projected/b683a409-b7d3-404f-ac5a-43be34e071c6-kube-api-access-mbmh5\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.086836 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/643ad5a5-986a-4003-a1ef-a0fc8a6cae10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.119329 4745 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 10 13:33:40 crc kubenswrapper[4745]: E1010 13:33:40.137376 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/horizon-66b896bb7f-64vcw" podUID="17718d89-5706-452c-a482-027b2c6f7c0d" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.143101 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "122e3abe-be44-46cc-826f-938bb985e0b2" (UID: "122e3abe-be44-46cc-826f-938bb985e0b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.166887 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b683a409-b7d3-404f-ac5a-43be34e071c6" (UID: "b683a409-b7d3-404f-ac5a-43be34e071c6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.172087 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-config-data" (OuterVolumeSpecName: "config-data") pod "122e3abe-be44-46cc-826f-938bb985e0b2" (UID: "122e3abe-be44-46cc-826f-938bb985e0b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.174289 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b683a409-b7d3-404f-ac5a-43be34e071c6" (UID: "b683a409-b7d3-404f-ac5a-43be34e071c6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.191278 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b683a409-b7d3-404f-ac5a-43be34e071c6" (UID: "b683a409-b7d3-404f-ac5a-43be34e071c6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.191931 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.191960 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.191974 4745 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.191983 4745 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.191991 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.192000 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122e3abe-be44-46cc-826f-938bb985e0b2-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.203322 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-config" (OuterVolumeSpecName: "config") pod "b683a409-b7d3-404f-ac5a-43be34e071c6" (UID: "b683a409-b7d3-404f-ac5a-43be34e071c6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.204259 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b683a409-b7d3-404f-ac5a-43be34e071c6" (UID: "b683a409-b7d3-404f-ac5a-43be34e071c6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:33:40 crc kubenswrapper[4745]: W1010 13:33:40.220802 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45009f2e_d0d5_4a20_b62c_d5d0534e1551.slice/crio-a85b3684c58b6faeef3d56c54eb1502b33619172c3103963b89eca08d70b6f4c WatchSource:0}: Error finding container a85b3684c58b6faeef3d56c54eb1502b33619172c3103963b89eca08d70b6f4c: Status 404 returned error can't find the container with id a85b3684c58b6faeef3d56c54eb1502b33619172c3103963b89eca08d70b6f4c Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.221458 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-krvls"] Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.240608 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-65c7d6f5bb-9g4xm"] Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.293764 4745 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.293792 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b683a409-b7d3-404f-ac5a-43be34e071c6-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.296346 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" event={"ID":"b683a409-b7d3-404f-ac5a-43be34e071c6","Type":"ContainerDied","Data":"234d83317618d819612a8f20c6629adb2a1d793bf4c92b3dfcb3119df6d0e7d2"} Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.296374 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-nv8xl" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.296395 4745 scope.go:117] "RemoveContainer" containerID="94760b0459376bca580d88cb72faa38ce3eb1269d08ffebd83ce22fcc54fb3ff" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.301206 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65c7d6f5bb-9g4xm" event={"ID":"604fc859-90dd-4c8b-a216-6009b3dbcbf2","Type":"ContainerStarted","Data":"32fb177c3021cf6bab2cafe5137093b0a0996835a356486b2465d7ed6093d116"} Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.304121 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66886d68fc-xt5ng" podUID="c688f752-0686-412c-85f8-8304cc245608" containerName="horizon" containerID="cri-o://e09f68f7592f8371de87ae35f055d708121a41f9294a944d4a80bbe2a529cbf7" gracePeriod=30 Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.304187 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66886d68fc-xt5ng" event={"ID":"c688f752-0686-412c-85f8-8304cc245608","Type":"ContainerStarted","Data":"e09f68f7592f8371de87ae35f055d708121a41f9294a944d4a80bbe2a529cbf7"} Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.329170 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"122e3abe-be44-46cc-826f-938bb985e0b2","Type":"ContainerDied","Data":"797444bcc9ca1460f95c0d36282a63ece76fc9799112b28b2ee945fd50856556"} Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.329255 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.329253 4745 scope.go:117] "RemoveContainer" containerID="4418d66aa17366f10075c640008b78b8e2f223f73a15dbd24ed2e4ace636096a" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.332042 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-krvls" event={"ID":"45009f2e-d0d5-4a20-b62c-d5d0534e1551","Type":"ContainerStarted","Data":"a85b3684c58b6faeef3d56c54eb1502b33619172c3103963b89eca08d70b6f4c"} Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.339983 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66b896bb7f-64vcw" event={"ID":"17718d89-5706-452c-a482-027b2c6f7c0d","Type":"ContainerStarted","Data":"25bb4181d3f192ce28d74d5196cad2fc3b503fc9146a353f9fb7043dbebeb9c1"} Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.340121 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66b896bb7f-64vcw" podUID="17718d89-5706-452c-a482-027b2c6f7c0d" containerName="horizon" containerID="cri-o://25bb4181d3f192ce28d74d5196cad2fc3b503fc9146a353f9fb7043dbebeb9c1" gracePeriod=30 Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.345129 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nwttm" event={"ID":"643ad5a5-986a-4003-a1ef-a0fc8a6cae10","Type":"ContainerDied","Data":"2b09594809465432cb7afe73a35b88e8005ec15306a31a7ec1716e0da7725ab4"} Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.345168 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b09594809465432cb7afe73a35b88e8005ec15306a31a7ec1716e0da7725ab4" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.345234 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nwttm" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.351067 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c875-account-create-mrv46" event={"ID":"a27645b6-74d3-4a15-ab07-c9767ef55413","Type":"ContainerDied","Data":"0b5f6b951c486eb09bf2cf9cba7f0719be4c498c01c38bab6f1ae08d55e7a6ab"} Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.351105 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b5f6b951c486eb09bf2cf9cba7f0719be4c498c01c38bab6f1ae08d55e7a6ab" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.351166 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c875-account-create-mrv46" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.354402 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"822d9514-5b19-440a-b4dc-5850f2bc8549","Type":"ContainerStarted","Data":"fb5d0bae521c81290ed4fc0fa35e521dba6cb81fc58a9fd2a68445497fdd0c2d"} Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.358975 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-nv8xl"] Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.359620 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-98a4-account-create-r4qqn" event={"ID":"235509a2-b516-432d-a681-9db08de1e7eb","Type":"ContainerDied","Data":"39fee32ea40d7584631052d739bab3ed07ed448905fd5515b3d23b2a606c893a"} Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.359709 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39fee32ea40d7584631052d739bab3ed07ed448905fd5515b3d23b2a606c893a" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.359800 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-98a4-account-create-r4qqn" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.364840 4745 scope.go:117] "RemoveContainer" containerID="da871c371d2c53e62ff85cc80535422abcceefce9f165cb531db25433f469db4" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.365544 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fcdc9c87c-9r2nf" event={"ID":"83abb158-a66f-48ed-ae55-40a2c35cfbf2","Type":"ContainerStarted","Data":"a7daaeb80ef807429fce0f697c35376cd88decd520f744be87f23e6c626826ae"} Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.365600 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fcdc9c87c-9r2nf" event={"ID":"83abb158-a66f-48ed-ae55-40a2c35cfbf2","Type":"ContainerStarted","Data":"35cd0b360583d80e4473257609b7bc4c69eaf342b6a359f2615b7e3491639440"} Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.365650 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6fcdc9c87c-9r2nf" podUID="83abb158-a66f-48ed-ae55-40a2c35cfbf2" containerName="horizon-log" containerID="cri-o://35cd0b360583d80e4473257609b7bc4c69eaf342b6a359f2615b7e3491639440" gracePeriod=30 Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.365675 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6fcdc9c87c-9r2nf" podUID="83abb158-a66f-48ed-ae55-40a2c35cfbf2" containerName="horizon" containerID="cri-o://a7daaeb80ef807429fce0f697c35376cd88decd520f744be87f23e6c626826ae" gracePeriod=30 Oct 10 13:33:40 crc kubenswrapper[4745]: E1010 13:33:40.369438 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-c2vrp" podUID="60ccdbf7-fc2d-4292-8ff6-d388ccda2b58" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.379436 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-nv8xl"] Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.408571 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8574589b48-pc9wr"] Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.424480 4745 scope.go:117] "RemoveContainer" containerID="935f60a55b4557fc3a9aa5989cd99b0947c2142f74f53ac34a86232fc8e2344f" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.442955 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6fcdc9c87c-9r2nf" podStartSLOduration=2.436846169 podStartE2EDuration="19.442932427s" podCreationTimestamp="2025-10-10 13:33:21 +0000 UTC" firstStartedPulling="2025-10-10 13:33:22.719308282 +0000 UTC m=+916.616965045" lastFinishedPulling="2025-10-10 13:33:39.72539454 +0000 UTC m=+933.623051303" observedRunningTime="2025-10-10 13:33:40.399967603 +0000 UTC m=+934.297624366" watchObservedRunningTime="2025-10-10 13:33:40.442932427 +0000 UTC m=+934.340589190" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.462798 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.486827 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.512787 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 13:33:40 crc kubenswrapper[4745]: E1010 13:33:40.513180 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="122e3abe-be44-46cc-826f-938bb985e0b2" containerName="glance-httpd" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.513196 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="122e3abe-be44-46cc-826f-938bb985e0b2" containerName="glance-httpd" Oct 10 13:33:40 crc kubenswrapper[4745]: E1010 13:33:40.513217 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b683a409-b7d3-404f-ac5a-43be34e071c6" containerName="init" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.513224 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="b683a409-b7d3-404f-ac5a-43be34e071c6" containerName="init" Oct 10 13:33:40 crc kubenswrapper[4745]: E1010 13:33:40.513232 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b683a409-b7d3-404f-ac5a-43be34e071c6" containerName="dnsmasq-dns" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.513238 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="b683a409-b7d3-404f-ac5a-43be34e071c6" containerName="dnsmasq-dns" Oct 10 13:33:40 crc kubenswrapper[4745]: E1010 13:33:40.513255 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27645b6-74d3-4a15-ab07-c9767ef55413" containerName="mariadb-account-create" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.513261 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27645b6-74d3-4a15-ab07-c9767ef55413" containerName="mariadb-account-create" Oct 10 13:33:40 crc kubenswrapper[4745]: E1010 13:33:40.513274 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="122e3abe-be44-46cc-826f-938bb985e0b2" containerName="glance-log" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.513279 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="122e3abe-be44-46cc-826f-938bb985e0b2" containerName="glance-log" Oct 10 13:33:40 crc kubenswrapper[4745]: E1010 13:33:40.513288 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="643ad5a5-986a-4003-a1ef-a0fc8a6cae10" containerName="keystone-bootstrap" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.513294 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="643ad5a5-986a-4003-a1ef-a0fc8a6cae10" containerName="keystone-bootstrap" Oct 10 13:33:40 crc kubenswrapper[4745]: E1010 13:33:40.513307 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="235509a2-b516-432d-a681-9db08de1e7eb" containerName="mariadb-account-create" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.513313 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="235509a2-b516-432d-a681-9db08de1e7eb" containerName="mariadb-account-create" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.513465 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="122e3abe-be44-46cc-826f-938bb985e0b2" containerName="glance-httpd" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.513476 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27645b6-74d3-4a15-ab07-c9767ef55413" containerName="mariadb-account-create" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.513487 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="643ad5a5-986a-4003-a1ef-a0fc8a6cae10" containerName="keystone-bootstrap" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.513499 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="235509a2-b516-432d-a681-9db08de1e7eb" containerName="mariadb-account-create" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.513517 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="122e3abe-be44-46cc-826f-938bb985e0b2" containerName="glance-log" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.513528 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="b683a409-b7d3-404f-ac5a-43be34e071c6" containerName="dnsmasq-dns" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.514419 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.525372 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.525785 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.531682 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.565449 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.599133 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btp2g\" (UniqueName: \"kubernetes.io/projected/ba6456ae-cc8a-45e4-8e01-d37a74e40722-kube-api-access-btp2g\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.600005 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.600028 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba6456ae-cc8a-45e4-8e01-d37a74e40722-logs\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.600044 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.600085 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.600101 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.600117 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.600166 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ba6456ae-cc8a-45e4-8e01-d37a74e40722-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.700565 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.700607 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.700623 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.700673 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ba6456ae-cc8a-45e4-8e01-d37a74e40722-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.700693 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btp2g\" (UniqueName: \"kubernetes.io/projected/ba6456ae-cc8a-45e4-8e01-d37a74e40722-kube-api-access-btp2g\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.700764 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba6456ae-cc8a-45e4-8e01-d37a74e40722-logs\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.700780 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.700794 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.701230 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ba6456ae-cc8a-45e4-8e01-d37a74e40722-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.701553 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba6456ae-cc8a-45e4-8e01-d37a74e40722-logs\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.702754 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.707700 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.707780 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.708279 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.708716 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.734183 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btp2g\" (UniqueName: \"kubernetes.io/projected/ba6456ae-cc8a-45e4-8e01-d37a74e40722-kube-api-access-btp2g\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.764842 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.769382 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="122e3abe-be44-46cc-826f-938bb985e0b2" path="/var/lib/kubelet/pods/122e3abe-be44-46cc-826f-938bb985e0b2/volumes" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.779714 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b683a409-b7d3-404f-ac5a-43be34e071c6" path="/var/lib/kubelet/pods/b683a409-b7d3-404f-ac5a-43be34e071c6/volumes" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.867989 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.970431 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-nwttm"] Oct 10 13:33:40 crc kubenswrapper[4745]: I1010 13:33:40.978061 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-nwttm"] Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.090413 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-p9wdk"] Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.091523 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.099868 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.100115 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.100221 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.100356 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r7tdf" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.112440 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-p9wdk"] Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.114358 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-combined-ca-bundle\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.114400 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-credential-keys\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.114421 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-fernet-keys\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.128276 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnnt5\" (UniqueName: \"kubernetes.io/projected/28786a80-f298-4618-8c54-5f943d96eff9-kube-api-access-vnnt5\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.128607 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-scripts\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.128633 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-config-data\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.233682 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnnt5\" (UniqueName: \"kubernetes.io/projected/28786a80-f298-4618-8c54-5f943d96eff9-kube-api-access-vnnt5\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.233780 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-scripts\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.233810 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-config-data\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.233840 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-combined-ca-bundle\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.233863 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-credential-keys\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.233889 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-fernet-keys\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.246914 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-config-data\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.247098 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-combined-ca-bundle\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.247406 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-fernet-keys\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.248125 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-scripts\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.252172 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-credential-keys\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.255913 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnnt5\" (UniqueName: \"kubernetes.io/projected/28786a80-f298-4618-8c54-5f943d96eff9-kube-api-access-vnnt5\") pod \"keystone-bootstrap-p9wdk\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.383489 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f","Type":"ContainerStarted","Data":"63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9"} Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.383776 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f","Type":"ContainerStarted","Data":"5a165eedb7127ac66919d490e50c12cf710a2ede38c9120d97db6f84c3e810e3"} Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.405935 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-krvls" event={"ID":"45009f2e-d0d5-4a20-b62c-d5d0534e1551","Type":"ContainerStarted","Data":"00bb0bf4c37ec198785551594973932ef355f5ed17da60e5fae2c842a5f44001"} Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.416824 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65c7d6f5bb-9g4xm" event={"ID":"604fc859-90dd-4c8b-a216-6009b3dbcbf2","Type":"ContainerStarted","Data":"078e3be7373635f32ef10c0d5768edb4110b552dd87f3f5a0f4b17afe3563529"} Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.416864 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65c7d6f5bb-9g4xm" event={"ID":"604fc859-90dd-4c8b-a216-6009b3dbcbf2","Type":"ContainerStarted","Data":"fec5f51d0b50e49ba472342af7f5e04b56f06dcd7a1d7afbd9ec19f6c386fb77"} Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.425386 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8574589b48-pc9wr" event={"ID":"fb40990e-1db6-4eab-9365-06678441a9ba","Type":"ContainerStarted","Data":"84af862f8e32632ffa9b7b3e3fc16b610948ac8ec958cfbc112015db35393892"} Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.425427 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8574589b48-pc9wr" event={"ID":"fb40990e-1db6-4eab-9365-06678441a9ba","Type":"ContainerStarted","Data":"93f27fe559f3760cfdc94b1b5e4b242ecc11a45162f94e60564ec71bb913126b"} Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.425438 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8574589b48-pc9wr" event={"ID":"fb40990e-1db6-4eab-9365-06678441a9ba","Type":"ContainerStarted","Data":"6939551919c89b068ab77e1e365aa010fda73aeee57cec63e1b4481b7392e841"} Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.431270 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.440910 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-krvls" podStartSLOduration=6.440891625 podStartE2EDuration="6.440891625s" podCreationTimestamp="2025-10-10 13:33:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:33:41.425036755 +0000 UTC m=+935.322693518" watchObservedRunningTime="2025-10-10 13:33:41.440891625 +0000 UTC m=+935.338548388" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.457079 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-65c7d6f5bb-9g4xm" podStartSLOduration=8.457061453 podStartE2EDuration="8.457061453s" podCreationTimestamp="2025-10-10 13:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:33:41.450304463 +0000 UTC m=+935.347961226" watchObservedRunningTime="2025-10-10 13:33:41.457061453 +0000 UTC m=+935.354718216" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.484339 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-8574589b48-pc9wr" podStartSLOduration=8.48431862 podStartE2EDuration="8.48431862s" podCreationTimestamp="2025-10-10 13:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:33:41.476144294 +0000 UTC m=+935.373801057" watchObservedRunningTime="2025-10-10 13:33:41.48431862 +0000 UTC m=+935.381975383" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.508874 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.822204 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.853814 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:33:41 crc kubenswrapper[4745]: I1010 13:33:41.955686 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-p9wdk"] Oct 10 13:33:42 crc kubenswrapper[4745]: I1010 13:33:42.442072 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ba6456ae-cc8a-45e4-8e01-d37a74e40722","Type":"ContainerStarted","Data":"103f7821fe3f99ff187f0620004c15f579d4c341b11b03269580e6565748be2a"} Oct 10 13:33:42 crc kubenswrapper[4745]: I1010 13:33:42.442114 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ba6456ae-cc8a-45e4-8e01-d37a74e40722","Type":"ContainerStarted","Data":"f43ce3d644d4b3802aaf005bebff396f83f23050492aaa9f4a5c76c2870c110b"} Oct 10 13:33:42 crc kubenswrapper[4745]: I1010 13:33:42.444297 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" containerName="glance-log" containerID="cri-o://63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9" gracePeriod=30 Oct 10 13:33:42 crc kubenswrapper[4745]: I1010 13:33:42.444565 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f","Type":"ContainerStarted","Data":"6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce"} Oct 10 13:33:42 crc kubenswrapper[4745]: I1010 13:33:42.445574 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" containerName="glance-httpd" containerID="cri-o://6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce" gracePeriod=30 Oct 10 13:33:42 crc kubenswrapper[4745]: I1010 13:33:42.801757 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="643ad5a5-986a-4003-a1ef-a0fc8a6cae10" path="/var/lib/kubelet/pods/643ad5a5-986a-4003-a1ef-a0fc8a6cae10/volumes" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.375497 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.411901 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.411951 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.476980 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"822d9514-5b19-440a-b4dc-5850f2bc8549","Type":"ContainerStarted","Data":"1772984a3fb04be75513af6fa054a11290ec24ca5f0b9560a2313d44f5b05cdd"} Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.484771 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ba6456ae-cc8a-45e4-8e01-d37a74e40722","Type":"ContainerStarted","Data":"fd34ad7ed94e880a266f45443e59d32a7e8ebfb7c590a7c3bac013fbc000b780"} Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.485403 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-scripts\") pod \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.485453 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-httpd-run\") pod \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.485511 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-logs\") pod \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.485552 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.485614 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tlc9\" (UniqueName: \"kubernetes.io/projected/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-kube-api-access-8tlc9\") pod \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.485685 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-combined-ca-bundle\") pod \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.485815 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-config-data\") pod \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\" (UID: \"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f\") " Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.486374 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-logs" (OuterVolumeSpecName: "logs") pod "5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" (UID: "5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.486549 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" (UID: "5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.487876 4745 generic.go:334] "Generic (PLEG): container finished" podID="5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" containerID="6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce" exitCode=0 Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.487900 4745 generic.go:334] "Generic (PLEG): container finished" podID="5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" containerID="63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9" exitCode=143 Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.487938 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f","Type":"ContainerDied","Data":"6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce"} Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.487962 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f","Type":"ContainerDied","Data":"63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9"} Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.487972 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f","Type":"ContainerDied","Data":"5a165eedb7127ac66919d490e50c12cf710a2ede38c9120d97db6f84c3e810e3"} Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.487986 4745 scope.go:117] "RemoveContainer" containerID="6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.488090 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.491451 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-kube-api-access-8tlc9" (OuterVolumeSpecName: "kube-api-access-8tlc9") pod "5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" (UID: "5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f"). InnerVolumeSpecName "kube-api-access-8tlc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.494271 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-scripts" (OuterVolumeSpecName: "scripts") pod "5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" (UID: "5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.497957 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p9wdk" event={"ID":"28786a80-f298-4618-8c54-5f943d96eff9","Type":"ContainerStarted","Data":"2feaf091557339dc922e1a4b1bd65e94052297f597c62ab49422faa9301c2315"} Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.497997 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p9wdk" event={"ID":"28786a80-f298-4618-8c54-5f943d96eff9","Type":"ContainerStarted","Data":"99f51eaccda9eed49e2cbc58b424d8bce5d145af1a22a56cdb684ebb67a8ebd5"} Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.511888 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" (UID: "5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.538943 4745 scope.go:117] "RemoveContainer" containerID="63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.539197 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.539182783 podStartE2EDuration="3.539182783s" podCreationTimestamp="2025-10-10 13:33:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:33:43.515141267 +0000 UTC m=+937.412798030" watchObservedRunningTime="2025-10-10 13:33:43.539182783 +0000 UTC m=+937.436839546" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.541195 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-p9wdk" podStartSLOduration=2.541188794 podStartE2EDuration="2.541188794s" podCreationTimestamp="2025-10-10 13:33:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:33:43.539531832 +0000 UTC m=+937.437188605" watchObservedRunningTime="2025-10-10 13:33:43.541188794 +0000 UTC m=+937.438845557" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.548715 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-config-data" (OuterVolumeSpecName: "config-data") pod "5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" (UID: "5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.564844 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" (UID: "5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.564877 4745 scope.go:117] "RemoveContainer" containerID="6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce" Oct 10 13:33:43 crc kubenswrapper[4745]: E1010 13:33:43.565583 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce\": container with ID starting with 6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce not found: ID does not exist" containerID="6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.565613 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce"} err="failed to get container status \"6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce\": rpc error: code = NotFound desc = could not find container \"6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce\": container with ID starting with 6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce not found: ID does not exist" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.565632 4745 scope.go:117] "RemoveContainer" containerID="63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9" Oct 10 13:33:43 crc kubenswrapper[4745]: E1010 13:33:43.568824 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9\": container with ID starting with 63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9 not found: ID does not exist" containerID="63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.568858 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9"} err="failed to get container status \"63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9\": rpc error: code = NotFound desc = could not find container \"63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9\": container with ID starting with 63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9 not found: ID does not exist" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.568877 4745 scope.go:117] "RemoveContainer" containerID="6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.570044 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce"} err="failed to get container status \"6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce\": rpc error: code = NotFound desc = could not find container \"6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce\": container with ID starting with 6a0911e3f3c7664982d84e8b08cd8d7df33bbe95579e74a6689730a875462dce not found: ID does not exist" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.570068 4745 scope.go:117] "RemoveContainer" containerID="63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.572055 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9"} err="failed to get container status \"63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9\": rpc error: code = NotFound desc = could not find container \"63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9\": container with ID starting with 63cb3ed3db97e03cb347a3691a22a3c348bb3967791a1ee04c4f711fee23a8c9 not found: ID does not exist" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.579861 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.579899 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.588475 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.588503 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.588512 4745 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.588529 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.588549 4745 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.588559 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tlc9\" (UniqueName: \"kubernetes.io/projected/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-kube-api-access-8tlc9\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.588569 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.604810 4745 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.690255 4745 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.828520 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.833577 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.866739 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:33:43 crc kubenswrapper[4745]: E1010 13:33:43.867054 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" containerName="glance-log" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.867083 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" containerName="glance-log" Oct 10 13:33:43 crc kubenswrapper[4745]: E1010 13:33:43.867113 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" containerName="glance-httpd" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.867120 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" containerName="glance-httpd" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.867287 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" containerName="glance-httpd" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.867313 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" containerName="glance-log" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.868118 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.870615 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.870771 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.939798 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.994573 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k4g9\" (UniqueName: \"kubernetes.io/projected/a0048c64-c35e-4358-8e8d-bdb0a65df21c-kube-api-access-9k4g9\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.994658 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.994723 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0048c64-c35e-4358-8e8d-bdb0a65df21c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.994885 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0048c64-c35e-4358-8e8d-bdb0a65df21c-logs\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.994989 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.995786 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.995836 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-scripts\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:43 crc kubenswrapper[4745]: I1010 13:33:43.995905 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-config-data\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.058431 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.097650 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0048c64-c35e-4358-8e8d-bdb0a65df21c-logs\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.098090 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0048c64-c35e-4358-8e8d-bdb0a65df21c-logs\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.098200 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.098824 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.098854 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-scripts\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.098888 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-config-data\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.099927 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k4g9\" (UniqueName: \"kubernetes.io/projected/a0048c64-c35e-4358-8e8d-bdb0a65df21c-kube-api-access-9k4g9\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.099995 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.100048 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0048c64-c35e-4358-8e8d-bdb0a65df21c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.100340 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0048c64-c35e-4358-8e8d-bdb0a65df21c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.100664 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.104373 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-scripts\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.105199 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.105794 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-config-data\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.112263 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.121537 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k4g9\" (UniqueName: \"kubernetes.io/projected/a0048c64-c35e-4358-8e8d-bdb0a65df21c-kube-api-access-9k4g9\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.132291 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.182609 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.758036 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f" path="/var/lib/kubelet/pods/5f8d9ab1-ccce-48fe-b1bd-9ffc054c508f/volumes" Oct 10 13:33:44 crc kubenswrapper[4745]: I1010 13:33:44.866150 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.534422 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0048c64-c35e-4358-8e8d-bdb0a65df21c","Type":"ContainerStarted","Data":"e2a93c49481809c48a060a8bc8aa9f7dea54c56006a4072561abb9ba51960103"} Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.639203 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-vskzt"] Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.660743 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vskzt"] Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.660833 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.665133 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wjzhr" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.665343 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.665511 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.748680 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-db-sync-config-data\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.748744 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-scripts\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.748770 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-config-data\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.748787 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-combined-ca-bundle\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.748851 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e86aa13-6741-4a98-88f7-1fec0870468f-etc-machine-id\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.748887 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ttxx\" (UniqueName: \"kubernetes.io/projected/7e86aa13-6741-4a98-88f7-1fec0870468f-kube-api-access-9ttxx\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.857916 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ttxx\" (UniqueName: \"kubernetes.io/projected/7e86aa13-6741-4a98-88f7-1fec0870468f-kube-api-access-9ttxx\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.858308 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-db-sync-config-data\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.858344 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-scripts\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.858370 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-config-data\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.858392 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-combined-ca-bundle\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.858461 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e86aa13-6741-4a98-88f7-1fec0870468f-etc-machine-id\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.858541 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e86aa13-6741-4a98-88f7-1fec0870468f-etc-machine-id\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.866301 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-scripts\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.867751 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-config-data\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.871566 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-db-sync-config-data\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.872346 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-combined-ca-bundle\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.881144 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ttxx\" (UniqueName: \"kubernetes.io/projected/7e86aa13-6741-4a98-88f7-1fec0870468f-kube-api-access-9ttxx\") pod \"cinder-db-sync-vskzt\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:45 crc kubenswrapper[4745]: I1010 13:33:45.977865 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vskzt" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.022790 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-dd6lk"] Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.027609 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dd6lk" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.030737 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-hw7zs" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.034314 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.041504 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dd6lk"] Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.174024 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7lh4\" (UniqueName: \"kubernetes.io/projected/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-kube-api-access-p7lh4\") pod \"barbican-db-sync-dd6lk\" (UID: \"7cbce4a4-b5ac-481c-87a7-47bc07e3361a\") " pod="openstack/barbican-db-sync-dd6lk" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.174103 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-db-sync-config-data\") pod \"barbican-db-sync-dd6lk\" (UID: \"7cbce4a4-b5ac-481c-87a7-47bc07e3361a\") " pod="openstack/barbican-db-sync-dd6lk" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.174141 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-combined-ca-bundle\") pod \"barbican-db-sync-dd6lk\" (UID: \"7cbce4a4-b5ac-481c-87a7-47bc07e3361a\") " pod="openstack/barbican-db-sync-dd6lk" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.188232 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.188283 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.188326 4745 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.188947 4745 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8a31c6076bf04e3e0b289d6ba4cb6b6d1eeb3e2ba416958cc06b859417161f1e"} pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.188998 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" containerID="cri-o://8a31c6076bf04e3e0b289d6ba4cb6b6d1eeb3e2ba416958cc06b859417161f1e" gracePeriod=600 Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.275828 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7lh4\" (UniqueName: \"kubernetes.io/projected/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-kube-api-access-p7lh4\") pod \"barbican-db-sync-dd6lk\" (UID: \"7cbce4a4-b5ac-481c-87a7-47bc07e3361a\") " pod="openstack/barbican-db-sync-dd6lk" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.275910 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-db-sync-config-data\") pod \"barbican-db-sync-dd6lk\" (UID: \"7cbce4a4-b5ac-481c-87a7-47bc07e3361a\") " pod="openstack/barbican-db-sync-dd6lk" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.275950 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-combined-ca-bundle\") pod \"barbican-db-sync-dd6lk\" (UID: \"7cbce4a4-b5ac-481c-87a7-47bc07e3361a\") " pod="openstack/barbican-db-sync-dd6lk" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.288723 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-combined-ca-bundle\") pod \"barbican-db-sync-dd6lk\" (UID: \"7cbce4a4-b5ac-481c-87a7-47bc07e3361a\") " pod="openstack/barbican-db-sync-dd6lk" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.291202 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-db-sync-config-data\") pod \"barbican-db-sync-dd6lk\" (UID: \"7cbce4a4-b5ac-481c-87a7-47bc07e3361a\") " pod="openstack/barbican-db-sync-dd6lk" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.324539 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7lh4\" (UniqueName: \"kubernetes.io/projected/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-kube-api-access-p7lh4\") pod \"barbican-db-sync-dd6lk\" (UID: \"7cbce4a4-b5ac-481c-87a7-47bc07e3361a\") " pod="openstack/barbican-db-sync-dd6lk" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.366269 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dd6lk" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.567250 4745 generic.go:334] "Generic (PLEG): container finished" podID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerID="8a31c6076bf04e3e0b289d6ba4cb6b6d1eeb3e2ba416958cc06b859417161f1e" exitCode=0 Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.567567 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerDied","Data":"8a31c6076bf04e3e0b289d6ba4cb6b6d1eeb3e2ba416958cc06b859417161f1e"} Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.567598 4745 scope.go:117] "RemoveContainer" containerID="2ba538b1feacb1e3fe2ff42d8dddb15b3e992a880af6f43548cdf890834090dc" Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.572541 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0048c64-c35e-4358-8e8d-bdb0a65df21c","Type":"ContainerStarted","Data":"9daf00e75c695335dcf8fcee3fbe2dc11a393a7a3845077d6f3cf5f5383fe27a"} Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.802079 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vskzt"] Oct 10 13:33:46 crc kubenswrapper[4745]: I1010 13:33:46.851618 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dd6lk"] Oct 10 13:33:47 crc kubenswrapper[4745]: I1010 13:33:47.622560 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"051be97974334031c215babb8b965fd2290d7b1b439e8dabb7659be3724a636f"} Oct 10 13:33:47 crc kubenswrapper[4745]: I1010 13:33:47.650490 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0048c64-c35e-4358-8e8d-bdb0a65df21c","Type":"ContainerStarted","Data":"17e01a5a08b518ada7b6fe6cf0842fb1a9ddeb2c4d93ec40c82f5f5354762385"} Oct 10 13:33:47 crc kubenswrapper[4745]: I1010 13:33:47.674390 4745 generic.go:334] "Generic (PLEG): container finished" podID="28786a80-f298-4618-8c54-5f943d96eff9" containerID="2feaf091557339dc922e1a4b1bd65e94052297f597c62ab49422faa9301c2315" exitCode=0 Oct 10 13:33:47 crc kubenswrapper[4745]: I1010 13:33:47.674689 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p9wdk" event={"ID":"28786a80-f298-4618-8c54-5f943d96eff9","Type":"ContainerDied","Data":"2feaf091557339dc922e1a4b1bd65e94052297f597c62ab49422faa9301c2315"} Oct 10 13:33:47 crc kubenswrapper[4745]: I1010 13:33:47.674801 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.674764601 podStartE2EDuration="4.674764601s" podCreationTimestamp="2025-10-10 13:33:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:33:47.673971701 +0000 UTC m=+941.571628464" watchObservedRunningTime="2025-10-10 13:33:47.674764601 +0000 UTC m=+941.572421364" Oct 10 13:33:47 crc kubenswrapper[4745]: I1010 13:33:47.679820 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vskzt" event={"ID":"7e86aa13-6741-4a98-88f7-1fec0870468f","Type":"ContainerStarted","Data":"c36b6d79fedb4cc7d59fd78cd5dee16082b1ffb26e970e4595bb6dd1f7d9b814"} Oct 10 13:33:47 crc kubenswrapper[4745]: I1010 13:33:47.682026 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dd6lk" event={"ID":"7cbce4a4-b5ac-481c-87a7-47bc07e3361a","Type":"ContainerStarted","Data":"aed9f254148e3d7868baf87730bca7496d68d58a7b3dcfb65e39cbab3eecaa37"} Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.330681 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.465547 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-scripts\") pod \"28786a80-f298-4618-8c54-5f943d96eff9\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.465674 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-config-data\") pod \"28786a80-f298-4618-8c54-5f943d96eff9\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.465712 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnnt5\" (UniqueName: \"kubernetes.io/projected/28786a80-f298-4618-8c54-5f943d96eff9-kube-api-access-vnnt5\") pod \"28786a80-f298-4618-8c54-5f943d96eff9\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.465754 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-credential-keys\") pod \"28786a80-f298-4618-8c54-5f943d96eff9\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.465832 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-combined-ca-bundle\") pod \"28786a80-f298-4618-8c54-5f943d96eff9\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.465895 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-fernet-keys\") pod \"28786a80-f298-4618-8c54-5f943d96eff9\" (UID: \"28786a80-f298-4618-8c54-5f943d96eff9\") " Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.479650 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28786a80-f298-4618-8c54-5f943d96eff9-kube-api-access-vnnt5" (OuterVolumeSpecName: "kube-api-access-vnnt5") pod "28786a80-f298-4618-8c54-5f943d96eff9" (UID: "28786a80-f298-4618-8c54-5f943d96eff9"). InnerVolumeSpecName "kube-api-access-vnnt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.487066 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-scripts" (OuterVolumeSpecName: "scripts") pod "28786a80-f298-4618-8c54-5f943d96eff9" (UID: "28786a80-f298-4618-8c54-5f943d96eff9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.491372 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "28786a80-f298-4618-8c54-5f943d96eff9" (UID: "28786a80-f298-4618-8c54-5f943d96eff9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.491399 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "28786a80-f298-4618-8c54-5f943d96eff9" (UID: "28786a80-f298-4618-8c54-5f943d96eff9"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.541999 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-config-data" (OuterVolumeSpecName: "config-data") pod "28786a80-f298-4618-8c54-5f943d96eff9" (UID: "28786a80-f298-4618-8c54-5f943d96eff9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.561855 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "28786a80-f298-4618-8c54-5f943d96eff9" (UID: "28786a80-f298-4618-8c54-5f943d96eff9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.570749 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.570790 4745 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.570801 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.570811 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.570819 4745 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/28786a80-f298-4618-8c54-5f943d96eff9-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.570828 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnnt5\" (UniqueName: \"kubernetes.io/projected/28786a80-f298-4618-8c54-5f943d96eff9-kube-api-access-vnnt5\") on node \"crc\" DevicePath \"\"" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.713803 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p9wdk" event={"ID":"28786a80-f298-4618-8c54-5f943d96eff9","Type":"ContainerDied","Data":"99f51eaccda9eed49e2cbc58b424d8bce5d145af1a22a56cdb684ebb67a8ebd5"} Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.713839 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99f51eaccda9eed49e2cbc58b424d8bce5d145af1a22a56cdb684ebb67a8ebd5" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.713893 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p9wdk" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.869103 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.869151 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.917672 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 10 13:33:50 crc kubenswrapper[4745]: I1010 13:33:50.921505 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.487128 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-745bfbd6-ms6d2"] Oct 10 13:33:51 crc kubenswrapper[4745]: E1010 13:33:51.487473 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28786a80-f298-4618-8c54-5f943d96eff9" containerName="keystone-bootstrap" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.487485 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="28786a80-f298-4618-8c54-5f943d96eff9" containerName="keystone-bootstrap" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.487655 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="28786a80-f298-4618-8c54-5f943d96eff9" containerName="keystone-bootstrap" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.489329 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.495186 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.495420 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.495570 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r7tdf" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.495681 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.495826 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.495919 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.501431 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-745bfbd6-ms6d2"] Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.593638 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8bs7\" (UniqueName: \"kubernetes.io/projected/a870bece-0e73-47f8-b0d2-79ee3e3cf637-kube-api-access-r8bs7\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.593691 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-credential-keys\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.593754 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-fernet-keys\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.593783 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-internal-tls-certs\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.593809 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-config-data\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.593860 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-combined-ca-bundle\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.594005 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-scripts\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.594078 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-public-tls-certs\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.695277 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8bs7\" (UniqueName: \"kubernetes.io/projected/a870bece-0e73-47f8-b0d2-79ee3e3cf637-kube-api-access-r8bs7\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.695326 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-credential-keys\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.695379 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-fernet-keys\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.695408 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-internal-tls-certs\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.695434 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-config-data\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.695465 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-combined-ca-bundle\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.695500 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-scripts\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.695523 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-public-tls-certs\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.699684 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-scripts\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.700744 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-config-data\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.701384 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-internal-tls-certs\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.702249 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-credential-keys\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.702578 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-public-tls-certs\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.702945 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-combined-ca-bundle\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.708968 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a870bece-0e73-47f8-b0d2-79ee3e3cf637-fernet-keys\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.710176 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8bs7\" (UniqueName: \"kubernetes.io/projected/a870bece-0e73-47f8-b0d2-79ee3e3cf637-kube-api-access-r8bs7\") pod \"keystone-745bfbd6-ms6d2\" (UID: \"a870bece-0e73-47f8-b0d2-79ee3e3cf637\") " pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.727624 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.728630 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 10 13:33:51 crc kubenswrapper[4745]: I1010 13:33:51.815451 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:33:53 crc kubenswrapper[4745]: I1010 13:33:53.414188 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-65c7d6f5bb-9g4xm" podUID="604fc859-90dd-4c8b-a216-6009b3dbcbf2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Oct 10 13:33:53 crc kubenswrapper[4745]: I1010 13:33:53.596671 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-8574589b48-pc9wr" podUID="fb40990e-1db6-4eab-9365-06678441a9ba" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 10 13:33:53 crc kubenswrapper[4745]: I1010 13:33:53.639615 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 10 13:33:53 crc kubenswrapper[4745]: I1010 13:33:53.744935 4745 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 10 13:33:54 crc kubenswrapper[4745]: I1010 13:33:54.184022 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 10 13:33:54 crc kubenswrapper[4745]: I1010 13:33:54.184303 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 10 13:33:54 crc kubenswrapper[4745]: I1010 13:33:54.216129 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 10 13:33:54 crc kubenswrapper[4745]: I1010 13:33:54.246564 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 10 13:33:54 crc kubenswrapper[4745]: I1010 13:33:54.307772 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 10 13:33:54 crc kubenswrapper[4745]: I1010 13:33:54.757946 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 10 13:33:54 crc kubenswrapper[4745]: I1010 13:33:54.757978 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 10 13:33:56 crc kubenswrapper[4745]: I1010 13:33:56.680144 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 10 13:33:56 crc kubenswrapper[4745]: I1010 13:33:56.764087 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 10 13:34:03 crc kubenswrapper[4745]: I1010 13:34:03.412305 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-65c7d6f5bb-9g4xm" podUID="604fc859-90dd-4c8b-a216-6009b3dbcbf2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Oct 10 13:34:04 crc kubenswrapper[4745]: I1010 13:34:04.855081 4745 generic.go:334] "Generic (PLEG): container finished" podID="45009f2e-d0d5-4a20-b62c-d5d0534e1551" containerID="00bb0bf4c37ec198785551594973932ef355f5ed17da60e5fae2c842a5f44001" exitCode=0 Oct 10 13:34:04 crc kubenswrapper[4745]: I1010 13:34:04.855175 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-krvls" event={"ID":"45009f2e-d0d5-4a20-b62c-d5d0534e1551","Type":"ContainerDied","Data":"00bb0bf4c37ec198785551594973932ef355f5ed17da60e5fae2c842a5f44001"} Oct 10 13:34:05 crc kubenswrapper[4745]: I1010 13:34:05.404074 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.002652 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-8574589b48-pc9wr" Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.089348 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-65c7d6f5bb-9g4xm"] Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.089875 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-65c7d6f5bb-9g4xm" podUID="604fc859-90dd-4c8b-a216-6009b3dbcbf2" containerName="horizon-log" containerID="cri-o://fec5f51d0b50e49ba472342af7f5e04b56f06dcd7a1d7afbd9ec19f6c386fb77" gracePeriod=30 Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.090310 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-65c7d6f5bb-9g4xm" podUID="604fc859-90dd-4c8b-a216-6009b3dbcbf2" containerName="horizon" containerID="cri-o://078e3be7373635f32ef10c0d5768edb4110b552dd87f3f5a0f4b17afe3563529" gracePeriod=30 Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.165008 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-krvls" Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.213867 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/45009f2e-d0d5-4a20-b62c-d5d0534e1551-config\") pod \"45009f2e-d0d5-4a20-b62c-d5d0534e1551\" (UID: \"45009f2e-d0d5-4a20-b62c-d5d0534e1551\") " Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.213989 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45009f2e-d0d5-4a20-b62c-d5d0534e1551-combined-ca-bundle\") pod \"45009f2e-d0d5-4a20-b62c-d5d0534e1551\" (UID: \"45009f2e-d0d5-4a20-b62c-d5d0534e1551\") " Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.214171 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wct7m\" (UniqueName: \"kubernetes.io/projected/45009f2e-d0d5-4a20-b62c-d5d0534e1551-kube-api-access-wct7m\") pod \"45009f2e-d0d5-4a20-b62c-d5d0534e1551\" (UID: \"45009f2e-d0d5-4a20-b62c-d5d0534e1551\") " Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.220147 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45009f2e-d0d5-4a20-b62c-d5d0534e1551-kube-api-access-wct7m" (OuterVolumeSpecName: "kube-api-access-wct7m") pod "45009f2e-d0d5-4a20-b62c-d5d0534e1551" (UID: "45009f2e-d0d5-4a20-b62c-d5d0534e1551"). InnerVolumeSpecName "kube-api-access-wct7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.246205 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45009f2e-d0d5-4a20-b62c-d5d0534e1551-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45009f2e-d0d5-4a20-b62c-d5d0534e1551" (UID: "45009f2e-d0d5-4a20-b62c-d5d0534e1551"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.264215 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45009f2e-d0d5-4a20-b62c-d5d0534e1551-config" (OuterVolumeSpecName: "config") pod "45009f2e-d0d5-4a20-b62c-d5d0534e1551" (UID: "45009f2e-d0d5-4a20-b62c-d5d0534e1551"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.316111 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/45009f2e-d0d5-4a20-b62c-d5d0534e1551-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.316141 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45009f2e-d0d5-4a20-b62c-d5d0534e1551-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.316152 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wct7m\" (UniqueName: \"kubernetes.io/projected/45009f2e-d0d5-4a20-b62c-d5d0534e1551-kube-api-access-wct7m\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.888380 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-krvls" event={"ID":"45009f2e-d0d5-4a20-b62c-d5d0534e1551","Type":"ContainerDied","Data":"a85b3684c58b6faeef3d56c54eb1502b33619172c3103963b89eca08d70b6f4c"} Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.888418 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a85b3684c58b6faeef3d56c54eb1502b33619172c3103963b89eca08d70b6f4c" Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.888475 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-krvls" Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.898899 4745 generic.go:334] "Generic (PLEG): container finished" podID="604fc859-90dd-4c8b-a216-6009b3dbcbf2" containerID="078e3be7373635f32ef10c0d5768edb4110b552dd87f3f5a0f4b17afe3563529" exitCode=0 Oct 10 13:34:07 crc kubenswrapper[4745]: I1010 13:34:07.899003 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65c7d6f5bb-9g4xm" event={"ID":"604fc859-90dd-4c8b-a216-6009b3dbcbf2","Type":"ContainerDied","Data":"078e3be7373635f32ef10c0d5768edb4110b552dd87f3f5a0f4b17afe3563529"} Oct 10 13:34:08 crc kubenswrapper[4745]: E1010 13:34:08.379797 4745 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 10 13:34:08 crc kubenswrapper[4745]: E1010 13:34:08.379965 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9ttxx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-vskzt_openstack(7e86aa13-6741-4a98-88f7-1fec0870468f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 13:34:08 crc kubenswrapper[4745]: E1010 13:34:08.382025 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-vskzt" podUID="7e86aa13-6741-4a98-88f7-1fec0870468f" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.432896 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-762hz"] Oct 10 13:34:08 crc kubenswrapper[4745]: E1010 13:34:08.433262 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45009f2e-d0d5-4a20-b62c-d5d0534e1551" containerName="neutron-db-sync" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.433275 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="45009f2e-d0d5-4a20-b62c-d5d0534e1551" containerName="neutron-db-sync" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.433488 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="45009f2e-d0d5-4a20-b62c-d5d0534e1551" containerName="neutron-db-sync" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.434450 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.451645 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-762hz"] Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.527212 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-748d64ff54-dltxm"] Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.528579 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.531868 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-wdk48" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.532923 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.533073 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.533233 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.549887 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.550120 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-528v8\" (UniqueName: \"kubernetes.io/projected/55673aa8-01ac-4a68-b059-cf1a1c0579d6-kube-api-access-528v8\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.550286 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.550304 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-config\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.550334 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.550363 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.565275 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-748d64ff54-dltxm"] Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.653555 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-528v8\" (UniqueName: \"kubernetes.io/projected/55673aa8-01ac-4a68-b059-cf1a1c0579d6-kube-api-access-528v8\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.653651 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-config\") pod \"neutron-748d64ff54-dltxm\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.653701 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-httpd-config\") pod \"neutron-748d64ff54-dltxm\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.653748 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkfhd\" (UniqueName: \"kubernetes.io/projected/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-kube-api-access-lkfhd\") pod \"neutron-748d64ff54-dltxm\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.653812 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.653844 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-config\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.653875 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-combined-ca-bundle\") pod \"neutron-748d64ff54-dltxm\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.653905 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.653937 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.654010 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-ovndb-tls-certs\") pod \"neutron-748d64ff54-dltxm\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.654050 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.655107 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.655621 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.657596 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.662238 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.666129 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-config\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.698537 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-528v8\" (UniqueName: \"kubernetes.io/projected/55673aa8-01ac-4a68-b059-cf1a1c0579d6-kube-api-access-528v8\") pod \"dnsmasq-dns-84b966f6c9-762hz\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.755452 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-config\") pod \"neutron-748d64ff54-dltxm\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.755794 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-httpd-config\") pod \"neutron-748d64ff54-dltxm\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.755817 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkfhd\" (UniqueName: \"kubernetes.io/projected/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-kube-api-access-lkfhd\") pod \"neutron-748d64ff54-dltxm\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.755866 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-combined-ca-bundle\") pod \"neutron-748d64ff54-dltxm\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.755922 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-ovndb-tls-certs\") pod \"neutron-748d64ff54-dltxm\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.761404 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-httpd-config\") pod \"neutron-748d64ff54-dltxm\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.766215 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-ovndb-tls-certs\") pod \"neutron-748d64ff54-dltxm\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.773064 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-combined-ca-bundle\") pod \"neutron-748d64ff54-dltxm\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.775498 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-config\") pod \"neutron-748d64ff54-dltxm\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.775906 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.802479 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkfhd\" (UniqueName: \"kubernetes.io/projected/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-kube-api-access-lkfhd\") pod \"neutron-748d64ff54-dltxm\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.931940 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-c2vrp" event={"ID":"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58","Type":"ContainerStarted","Data":"262bb16c312dd05eb8a87ac07d05f484e890cc5a2a2441b648b10260e103fa6e"} Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.940197 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dd6lk" event={"ID":"7cbce4a4-b5ac-481c-87a7-47bc07e3361a","Type":"ContainerStarted","Data":"63a78cd681c0ab49c65e8b0702691600a179cf2b4f4e6340c54a309ca98cddd9"} Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.940929 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.951657 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"822d9514-5b19-440a-b4dc-5850f2bc8549","Type":"ContainerStarted","Data":"4695a67f02f9e358f724e4f6181357e4d97926fa6e352f01577f0d646306bddc"} Oct 10 13:34:08 crc kubenswrapper[4745]: E1010 13:34:08.957179 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-vskzt" podUID="7e86aa13-6741-4a98-88f7-1fec0870468f" Oct 10 13:34:08 crc kubenswrapper[4745]: I1010 13:34:08.958813 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-c2vrp" podStartSLOduration=2.331733428 podStartE2EDuration="47.958795139s" podCreationTimestamp="2025-10-10 13:33:21 +0000 UTC" firstStartedPulling="2025-10-10 13:33:22.912045453 +0000 UTC m=+916.809702216" lastFinishedPulling="2025-10-10 13:34:08.539107164 +0000 UTC m=+962.436763927" observedRunningTime="2025-10-10 13:34:08.951387902 +0000 UTC m=+962.849044685" watchObservedRunningTime="2025-10-10 13:34:08.958795139 +0000 UTC m=+962.856451892" Oct 10 13:34:09 crc kubenswrapper[4745]: I1010 13:34:09.016583 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-dd6lk" podStartSLOduration=2.351230442 podStartE2EDuration="24.016565316s" podCreationTimestamp="2025-10-10 13:33:45 +0000 UTC" firstStartedPulling="2025-10-10 13:33:46.874793126 +0000 UTC m=+940.772449889" lastFinishedPulling="2025-10-10 13:34:08.540128 +0000 UTC m=+962.437784763" observedRunningTime="2025-10-10 13:34:09.006588034 +0000 UTC m=+962.904244797" watchObservedRunningTime="2025-10-10 13:34:09.016565316 +0000 UTC m=+962.914222079" Oct 10 13:34:09 crc kubenswrapper[4745]: I1010 13:34:09.054990 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-745bfbd6-ms6d2"] Oct 10 13:34:09 crc kubenswrapper[4745]: I1010 13:34:09.268919 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-762hz"] Oct 10 13:34:09 crc kubenswrapper[4745]: I1010 13:34:09.807194 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-748d64ff54-dltxm"] Oct 10 13:34:09 crc kubenswrapper[4745]: I1010 13:34:09.958502 4745 generic.go:334] "Generic (PLEG): container finished" podID="55673aa8-01ac-4a68-b059-cf1a1c0579d6" containerID="9f5fb23de37bb84f17c9eac9238bce93713ed1ab457df225d4dffde8aef9657c" exitCode=0 Oct 10 13:34:09 crc kubenswrapper[4745]: I1010 13:34:09.958557 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-762hz" event={"ID":"55673aa8-01ac-4a68-b059-cf1a1c0579d6","Type":"ContainerDied","Data":"9f5fb23de37bb84f17c9eac9238bce93713ed1ab457df225d4dffde8aef9657c"} Oct 10 13:34:09 crc kubenswrapper[4745]: I1010 13:34:09.958582 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-762hz" event={"ID":"55673aa8-01ac-4a68-b059-cf1a1c0579d6","Type":"ContainerStarted","Data":"576a1186a5f0b7e385429ac91d4ed7d006654896d3f44fdb7c265538a2c25809"} Oct 10 13:34:09 crc kubenswrapper[4745]: I1010 13:34:09.961197 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-748d64ff54-dltxm" event={"ID":"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb","Type":"ContainerStarted","Data":"8d562c7487b56dce124c35e8e9669ce372d339a4efb7c6150cd4883889b75766"} Oct 10 13:34:09 crc kubenswrapper[4745]: I1010 13:34:09.964718 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-745bfbd6-ms6d2" event={"ID":"a870bece-0e73-47f8-b0d2-79ee3e3cf637","Type":"ContainerStarted","Data":"2b939dfc27af9d3d8cee9f62f2d792667bfc3a991c3069ff20bbd14dfd52ce53"} Oct 10 13:34:09 crc kubenswrapper[4745]: I1010 13:34:09.964785 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-745bfbd6-ms6d2" event={"ID":"a870bece-0e73-47f8-b0d2-79ee3e3cf637","Type":"ContainerStarted","Data":"d8bf3a0c3e2144793d99bad5b0dc004d05a9f7a03c41726a7984cbd433b08a8a"} Oct 10 13:34:09 crc kubenswrapper[4745]: I1010 13:34:09.964819 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.586177 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-745bfbd6-ms6d2" podStartSLOduration=19.58612832 podStartE2EDuration="19.58612832s" podCreationTimestamp="2025-10-10 13:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:34:10.002834289 +0000 UTC m=+963.900491062" watchObservedRunningTime="2025-10-10 13:34:10.58612832 +0000 UTC m=+964.483785093" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.587632 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-86b9f849df-zktvk"] Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.591342 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.597998 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.598171 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.599323 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-86b9f849df-zktvk"] Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.700636 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-internal-tls-certs\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.700691 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92zjx\" (UniqueName: \"kubernetes.io/projected/668e04d2-c914-4325-acab-c8638172e4cc-kube-api-access-92zjx\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.700751 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-public-tls-certs\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.700778 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-httpd-config\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.700804 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-config\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.700843 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-combined-ca-bundle\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.700862 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-ovndb-tls-certs\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.804634 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-internal-tls-certs\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.804689 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92zjx\" (UniqueName: \"kubernetes.io/projected/668e04d2-c914-4325-acab-c8638172e4cc-kube-api-access-92zjx\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.804753 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-public-tls-certs\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.804777 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-httpd-config\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.804803 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-config\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.818935 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-combined-ca-bundle\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.818968 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-ovndb-tls-certs\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.810670 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-config\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.823213 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-public-tls-certs\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.825396 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-ovndb-tls-certs\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.837402 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-internal-tls-certs\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.838048 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92zjx\" (UniqueName: \"kubernetes.io/projected/668e04d2-c914-4325-acab-c8638172e4cc-kube-api-access-92zjx\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.858004 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-httpd-config\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.860922 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/668e04d2-c914-4325-acab-c8638172e4cc-combined-ca-bundle\") pod \"neutron-86b9f849df-zktvk\" (UID: \"668e04d2-c914-4325-acab-c8638172e4cc\") " pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.973277 4745 generic.go:334] "Generic (PLEG): container finished" podID="83abb158-a66f-48ed-ae55-40a2c35cfbf2" containerID="a7daaeb80ef807429fce0f697c35376cd88decd520f744be87f23e6c626826ae" exitCode=137 Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.973323 4745 generic.go:334] "Generic (PLEG): container finished" podID="83abb158-a66f-48ed-ae55-40a2c35cfbf2" containerID="35cd0b360583d80e4473257609b7bc4c69eaf342b6a359f2615b7e3491639440" exitCode=137 Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.973389 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fcdc9c87c-9r2nf" event={"ID":"83abb158-a66f-48ed-ae55-40a2c35cfbf2","Type":"ContainerDied","Data":"a7daaeb80ef807429fce0f697c35376cd88decd520f744be87f23e6c626826ae"} Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.973418 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fcdc9c87c-9r2nf" event={"ID":"83abb158-a66f-48ed-ae55-40a2c35cfbf2","Type":"ContainerDied","Data":"35cd0b360583d80e4473257609b7bc4c69eaf342b6a359f2615b7e3491639440"} Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.979275 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-762hz" event={"ID":"55673aa8-01ac-4a68-b059-cf1a1c0579d6","Type":"ContainerStarted","Data":"e745532a6a236899e940920e329204fed2f608738e54734da0e920f0a4765c9e"} Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.979770 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.986900 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-748d64ff54-dltxm" event={"ID":"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb","Type":"ContainerStarted","Data":"f7973f4a6d00df3d2f32ea7cc7625ef1ee20cee5b1a1d62752a024768b496050"} Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.986940 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-748d64ff54-dltxm" event={"ID":"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb","Type":"ContainerStarted","Data":"9376a83e2ad2b48814abda7d320b0ce3611a5cba3cda60126cab6cee90ecd747"} Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.987152 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.990156 4745 generic.go:334] "Generic (PLEG): container finished" podID="c688f752-0686-412c-85f8-8304cc245608" containerID="e09f68f7592f8371de87ae35f055d708121a41f9294a944d4a80bbe2a529cbf7" exitCode=137 Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.990215 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66886d68fc-xt5ng" event={"ID":"c688f752-0686-412c-85f8-8304cc245608","Type":"ContainerDied","Data":"e09f68f7592f8371de87ae35f055d708121a41f9294a944d4a80bbe2a529cbf7"} Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.991165 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.992465 4745 generic.go:334] "Generic (PLEG): container finished" podID="17718d89-5706-452c-a482-027b2c6f7c0d" containerID="25bb4181d3f192ce28d74d5196cad2fc3b503fc9146a353f9fb7043dbebeb9c1" exitCode=137 Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.992983 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66b896bb7f-64vcw" event={"ID":"17718d89-5706-452c-a482-027b2c6f7c0d","Type":"ContainerDied","Data":"25bb4181d3f192ce28d74d5196cad2fc3b503fc9146a353f9fb7043dbebeb9c1"} Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.993044 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66b896bb7f-64vcw" event={"ID":"17718d89-5706-452c-a482-027b2c6f7c0d","Type":"ContainerDied","Data":"6bc8a5904b5bb03d069ec0a58dc80f185db938ce6f84821a384b5d5b4aa771de"} Oct 10 13:34:10 crc kubenswrapper[4745]: I1010 13:34:10.993071 4745 scope.go:117] "RemoveContainer" containerID="25bb4181d3f192ce28d74d5196cad2fc3b503fc9146a353f9fb7043dbebeb9c1" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.012255 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b966f6c9-762hz" podStartSLOduration=3.012232445 podStartE2EDuration="3.012232445s" podCreationTimestamp="2025-10-10 13:34:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:34:11.006815239 +0000 UTC m=+964.904472012" watchObservedRunningTime="2025-10-10 13:34:11.012232445 +0000 UTC m=+964.909889198" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.047372 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.057497 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-748d64ff54-dltxm" podStartSLOduration=3.057475496 podStartE2EDuration="3.057475496s" podCreationTimestamp="2025-10-10 13:34:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:34:11.047048483 +0000 UTC m=+964.944705236" watchObservedRunningTime="2025-10-10 13:34:11.057475496 +0000 UTC m=+964.955132259" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.125258 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17718d89-5706-452c-a482-027b2c6f7c0d-config-data\") pod \"17718d89-5706-452c-a482-027b2c6f7c0d\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.125818 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj2lz\" (UniqueName: \"kubernetes.io/projected/17718d89-5706-452c-a482-027b2c6f7c0d-kube-api-access-mj2lz\") pod \"17718d89-5706-452c-a482-027b2c6f7c0d\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.125877 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17718d89-5706-452c-a482-027b2c6f7c0d-horizon-secret-key\") pod \"17718d89-5706-452c-a482-027b2c6f7c0d\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.125962 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17718d89-5706-452c-a482-027b2c6f7c0d-logs\") pod \"17718d89-5706-452c-a482-027b2c6f7c0d\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.126021 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17718d89-5706-452c-a482-027b2c6f7c0d-scripts\") pod \"17718d89-5706-452c-a482-027b2c6f7c0d\" (UID: \"17718d89-5706-452c-a482-027b2c6f7c0d\") " Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.130896 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17718d89-5706-452c-a482-027b2c6f7c0d-kube-api-access-mj2lz" (OuterVolumeSpecName: "kube-api-access-mj2lz") pod "17718d89-5706-452c-a482-027b2c6f7c0d" (UID: "17718d89-5706-452c-a482-027b2c6f7c0d"). InnerVolumeSpecName "kube-api-access-mj2lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.132592 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17718d89-5706-452c-a482-027b2c6f7c0d-logs" (OuterVolumeSpecName: "logs") pod "17718d89-5706-452c-a482-027b2c6f7c0d" (UID: "17718d89-5706-452c-a482-027b2c6f7c0d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.134173 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17718d89-5706-452c-a482-027b2c6f7c0d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "17718d89-5706-452c-a482-027b2c6f7c0d" (UID: "17718d89-5706-452c-a482-027b2c6f7c0d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.149542 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17718d89-5706-452c-a482-027b2c6f7c0d-config-data" (OuterVolumeSpecName: "config-data") pod "17718d89-5706-452c-a482-027b2c6f7c0d" (UID: "17718d89-5706-452c-a482-027b2c6f7c0d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.167038 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17718d89-5706-452c-a482-027b2c6f7c0d-scripts" (OuterVolumeSpecName: "scripts") pod "17718d89-5706-452c-a482-027b2c6f7c0d" (UID: "17718d89-5706-452c-a482-027b2c6f7c0d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.227870 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj2lz\" (UniqueName: \"kubernetes.io/projected/17718d89-5706-452c-a482-027b2c6f7c0d-kube-api-access-mj2lz\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.227909 4745 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17718d89-5706-452c-a482-027b2c6f7c0d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.227922 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17718d89-5706-452c-a482-027b2c6f7c0d-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.227934 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17718d89-5706-452c-a482-027b2c6f7c0d-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.227944 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17718d89-5706-452c-a482-027b2c6f7c0d-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.235310 4745 scope.go:117] "RemoveContainer" containerID="25bb4181d3f192ce28d74d5196cad2fc3b503fc9146a353f9fb7043dbebeb9c1" Oct 10 13:34:11 crc kubenswrapper[4745]: E1010 13:34:11.235681 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25bb4181d3f192ce28d74d5196cad2fc3b503fc9146a353f9fb7043dbebeb9c1\": container with ID starting with 25bb4181d3f192ce28d74d5196cad2fc3b503fc9146a353f9fb7043dbebeb9c1 not found: ID does not exist" containerID="25bb4181d3f192ce28d74d5196cad2fc3b503fc9146a353f9fb7043dbebeb9c1" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.235712 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25bb4181d3f192ce28d74d5196cad2fc3b503fc9146a353f9fb7043dbebeb9c1"} err="failed to get container status \"25bb4181d3f192ce28d74d5196cad2fc3b503fc9146a353f9fb7043dbebeb9c1\": rpc error: code = NotFound desc = could not find container \"25bb4181d3f192ce28d74d5196cad2fc3b503fc9146a353f9fb7043dbebeb9c1\": container with ID starting with 25bb4181d3f192ce28d74d5196cad2fc3b503fc9146a353f9fb7043dbebeb9c1 not found: ID does not exist" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.252690 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.254534 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.328760 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/83abb158-a66f-48ed-ae55-40a2c35cfbf2-config-data\") pod \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.328825 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83abb158-a66f-48ed-ae55-40a2c35cfbf2-scripts\") pod \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.328857 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/83abb158-a66f-48ed-ae55-40a2c35cfbf2-horizon-secret-key\") pod \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.328911 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c688f752-0686-412c-85f8-8304cc245608-logs\") pod \"c688f752-0686-412c-85f8-8304cc245608\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.328989 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c688f752-0686-412c-85f8-8304cc245608-scripts\") pod \"c688f752-0686-412c-85f8-8304cc245608\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.329021 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vlfx\" (UniqueName: \"kubernetes.io/projected/c688f752-0686-412c-85f8-8304cc245608-kube-api-access-9vlfx\") pod \"c688f752-0686-412c-85f8-8304cc245608\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.329128 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gv94\" (UniqueName: \"kubernetes.io/projected/83abb158-a66f-48ed-ae55-40a2c35cfbf2-kube-api-access-2gv94\") pod \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.329168 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83abb158-a66f-48ed-ae55-40a2c35cfbf2-logs\") pod \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\" (UID: \"83abb158-a66f-48ed-ae55-40a2c35cfbf2\") " Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.329196 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c688f752-0686-412c-85f8-8304cc245608-config-data\") pod \"c688f752-0686-412c-85f8-8304cc245608\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.329226 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c688f752-0686-412c-85f8-8304cc245608-horizon-secret-key\") pod \"c688f752-0686-412c-85f8-8304cc245608\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.331075 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c688f752-0686-412c-85f8-8304cc245608-logs" (OuterVolumeSpecName: "logs") pod "c688f752-0686-412c-85f8-8304cc245608" (UID: "c688f752-0686-412c-85f8-8304cc245608"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.331531 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83abb158-a66f-48ed-ae55-40a2c35cfbf2-logs" (OuterVolumeSpecName: "logs") pod "83abb158-a66f-48ed-ae55-40a2c35cfbf2" (UID: "83abb158-a66f-48ed-ae55-40a2c35cfbf2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.337062 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83abb158-a66f-48ed-ae55-40a2c35cfbf2-kube-api-access-2gv94" (OuterVolumeSpecName: "kube-api-access-2gv94") pod "83abb158-a66f-48ed-ae55-40a2c35cfbf2" (UID: "83abb158-a66f-48ed-ae55-40a2c35cfbf2"). InnerVolumeSpecName "kube-api-access-2gv94". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.337062 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83abb158-a66f-48ed-ae55-40a2c35cfbf2-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "83abb158-a66f-48ed-ae55-40a2c35cfbf2" (UID: "83abb158-a66f-48ed-ae55-40a2c35cfbf2"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.337053 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c688f752-0686-412c-85f8-8304cc245608-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c688f752-0686-412c-85f8-8304cc245608" (UID: "c688f752-0686-412c-85f8-8304cc245608"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.337209 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c688f752-0686-412c-85f8-8304cc245608-kube-api-access-9vlfx" (OuterVolumeSpecName: "kube-api-access-9vlfx") pod "c688f752-0686-412c-85f8-8304cc245608" (UID: "c688f752-0686-412c-85f8-8304cc245608"). InnerVolumeSpecName "kube-api-access-9vlfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:11 crc kubenswrapper[4745]: E1010 13:34:11.353049 4745 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c688f752-0686-412c-85f8-8304cc245608-scripts podName:c688f752-0686-412c-85f8-8304cc245608 nodeName:}" failed. No retries permitted until 2025-10-10 13:34:11.85301944 +0000 UTC m=+965.750676203 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "scripts" (UniqueName: "kubernetes.io/configmap/c688f752-0686-412c-85f8-8304cc245608-scripts") pod "c688f752-0686-412c-85f8-8304cc245608" (UID: "c688f752-0686-412c-85f8-8304cc245608") : error deleting /var/lib/kubelet/pods/c688f752-0686-412c-85f8-8304cc245608/volume-subpaths: remove /var/lib/kubelet/pods/c688f752-0686-412c-85f8-8304cc245608/volume-subpaths: no such file or directory Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.353716 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c688f752-0686-412c-85f8-8304cc245608-config-data" (OuterVolumeSpecName: "config-data") pod "c688f752-0686-412c-85f8-8304cc245608" (UID: "c688f752-0686-412c-85f8-8304cc245608"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.353758 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83abb158-a66f-48ed-ae55-40a2c35cfbf2-config-data" (OuterVolumeSpecName: "config-data") pod "83abb158-a66f-48ed-ae55-40a2c35cfbf2" (UID: "83abb158-a66f-48ed-ae55-40a2c35cfbf2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.355597 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83abb158-a66f-48ed-ae55-40a2c35cfbf2-scripts" (OuterVolumeSpecName: "scripts") pod "83abb158-a66f-48ed-ae55-40a2c35cfbf2" (UID: "83abb158-a66f-48ed-ae55-40a2c35cfbf2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.430834 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gv94\" (UniqueName: \"kubernetes.io/projected/83abb158-a66f-48ed-ae55-40a2c35cfbf2-kube-api-access-2gv94\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.430862 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83abb158-a66f-48ed-ae55-40a2c35cfbf2-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.430872 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c688f752-0686-412c-85f8-8304cc245608-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.430880 4745 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c688f752-0686-412c-85f8-8304cc245608-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.430889 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/83abb158-a66f-48ed-ae55-40a2c35cfbf2-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.430897 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83abb158-a66f-48ed-ae55-40a2c35cfbf2-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.430905 4745 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/83abb158-a66f-48ed-ae55-40a2c35cfbf2-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.430914 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c688f752-0686-412c-85f8-8304cc245608-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.430922 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vlfx\" (UniqueName: \"kubernetes.io/projected/c688f752-0686-412c-85f8-8304cc245608-kube-api-access-9vlfx\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.809638 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-86b9f849df-zktvk"] Oct 10 13:34:11 crc kubenswrapper[4745]: W1010 13:34:11.813922 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod668e04d2_c914_4325_acab_c8638172e4cc.slice/crio-46602bced20fd4542a21a66a94092139a3e660d4635a426ad7dfdf686f0cb89f WatchSource:0}: Error finding container 46602bced20fd4542a21a66a94092139a3e660d4635a426ad7dfdf686f0cb89f: Status 404 returned error can't find the container with id 46602bced20fd4542a21a66a94092139a3e660d4635a426ad7dfdf686f0cb89f Oct 10 13:34:11 crc kubenswrapper[4745]: E1010 13:34:11.830599 4745 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60ccdbf7_fc2d_4292_8ff6_d388ccda2b58.slice/crio-conmon-262bb16c312dd05eb8a87ac07d05f484e890cc5a2a2441b648b10260e103fa6e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60ccdbf7_fc2d_4292_8ff6_d388ccda2b58.slice/crio-262bb16c312dd05eb8a87ac07d05f484e890cc5a2a2441b648b10260e103fa6e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45009f2e_d0d5_4a20_b62c_d5d0534e1551.slice/crio-a85b3684c58b6faeef3d56c54eb1502b33619172c3103963b89eca08d70b6f4c\": RecentStats: unable to find data in memory cache]" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.942771 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c688f752-0686-412c-85f8-8304cc245608-scripts\") pod \"c688f752-0686-412c-85f8-8304cc245608\" (UID: \"c688f752-0686-412c-85f8-8304cc245608\") " Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.943493 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c688f752-0686-412c-85f8-8304cc245608-scripts" (OuterVolumeSpecName: "scripts") pod "c688f752-0686-412c-85f8-8304cc245608" (UID: "c688f752-0686-412c-85f8-8304cc245608"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:11 crc kubenswrapper[4745]: I1010 13:34:11.944828 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c688f752-0686-412c-85f8-8304cc245608-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.005826 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86b9f849df-zktvk" event={"ID":"668e04d2-c914-4325-acab-c8638172e4cc","Type":"ContainerStarted","Data":"46602bced20fd4542a21a66a94092139a3e660d4635a426ad7dfdf686f0cb89f"} Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.008394 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66886d68fc-xt5ng" Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.008374 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66886d68fc-xt5ng" event={"ID":"c688f752-0686-412c-85f8-8304cc245608","Type":"ContainerDied","Data":"b38149694aab57e091d93d19e32a6fd16eafa00dd1161a676771b76b70dfb84e"} Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.008544 4745 scope.go:117] "RemoveContainer" containerID="e09f68f7592f8371de87ae35f055d708121a41f9294a944d4a80bbe2a529cbf7" Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.010768 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66b896bb7f-64vcw" Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.020945 4745 generic.go:334] "Generic (PLEG): container finished" podID="60ccdbf7-fc2d-4292-8ff6-d388ccda2b58" containerID="262bb16c312dd05eb8a87ac07d05f484e890cc5a2a2441b648b10260e103fa6e" exitCode=0 Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.021007 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-c2vrp" event={"ID":"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58","Type":"ContainerDied","Data":"262bb16c312dd05eb8a87ac07d05f484e890cc5a2a2441b648b10260e103fa6e"} Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.025666 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fcdc9c87c-9r2nf" event={"ID":"83abb158-a66f-48ed-ae55-40a2c35cfbf2","Type":"ContainerDied","Data":"5486e6e9d0a18e48d94f08e3c2f3f6dfb047e1ba2f47d1785d8b378fe3dc6de8"} Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.025945 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fcdc9c87c-9r2nf" Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.119136 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66886d68fc-xt5ng"] Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.130354 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-66886d68fc-xt5ng"] Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.151254 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66b896bb7f-64vcw"] Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.157183 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-66b896bb7f-64vcw"] Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.163830 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6fcdc9c87c-9r2nf"] Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.170664 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6fcdc9c87c-9r2nf"] Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.228956 4745 scope.go:117] "RemoveContainer" containerID="a7daaeb80ef807429fce0f697c35376cd88decd520f744be87f23e6c626826ae" Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.422066 4745 scope.go:117] "RemoveContainer" containerID="35cd0b360583d80e4473257609b7bc4c69eaf342b6a359f2615b7e3491639440" Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.764431 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17718d89-5706-452c-a482-027b2c6f7c0d" path="/var/lib/kubelet/pods/17718d89-5706-452c-a482-027b2c6f7c0d/volumes" Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.766013 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83abb158-a66f-48ed-ae55-40a2c35cfbf2" path="/var/lib/kubelet/pods/83abb158-a66f-48ed-ae55-40a2c35cfbf2/volumes" Oct 10 13:34:12 crc kubenswrapper[4745]: I1010 13:34:12.767442 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c688f752-0686-412c-85f8-8304cc245608" path="/var/lib/kubelet/pods/c688f752-0686-412c-85f8-8304cc245608/volumes" Oct 10 13:34:13 crc kubenswrapper[4745]: I1010 13:34:13.035985 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86b9f849df-zktvk" event={"ID":"668e04d2-c914-4325-acab-c8638172e4cc","Type":"ContainerStarted","Data":"1167b781c2d338947baae7baeb10d5f12a0f26735ac1c9762958a88f01229552"} Oct 10 13:34:13 crc kubenswrapper[4745]: I1010 13:34:13.036024 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86b9f849df-zktvk" event={"ID":"668e04d2-c914-4325-acab-c8638172e4cc","Type":"ContainerStarted","Data":"0b1bab695046897bf602b81781550def6b3c723cd0e59d12339fa3cb49920db1"} Oct 10 13:34:13 crc kubenswrapper[4745]: I1010 13:34:13.038687 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:13 crc kubenswrapper[4745]: I1010 13:34:13.041546 4745 generic.go:334] "Generic (PLEG): container finished" podID="7cbce4a4-b5ac-481c-87a7-47bc07e3361a" containerID="63a78cd681c0ab49c65e8b0702691600a179cf2b4f4e6340c54a309ca98cddd9" exitCode=0 Oct 10 13:34:13 crc kubenswrapper[4745]: I1010 13:34:13.041916 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dd6lk" event={"ID":"7cbce4a4-b5ac-481c-87a7-47bc07e3361a","Type":"ContainerDied","Data":"63a78cd681c0ab49c65e8b0702691600a179cf2b4f4e6340c54a309ca98cddd9"} Oct 10 13:34:13 crc kubenswrapper[4745]: I1010 13:34:13.081048 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-86b9f849df-zktvk" podStartSLOduration=3.0810234 podStartE2EDuration="3.0810234s" podCreationTimestamp="2025-10-10 13:34:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:34:13.061236841 +0000 UTC m=+966.958893614" watchObservedRunningTime="2025-10-10 13:34:13.0810234 +0000 UTC m=+966.978680173" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.478614 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-c2vrp" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.482036 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dd6lk" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.581697 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-config-data\") pod \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.581789 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-db-sync-config-data\") pod \"7cbce4a4-b5ac-481c-87a7-47bc07e3361a\" (UID: \"7cbce4a4-b5ac-481c-87a7-47bc07e3361a\") " Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.581874 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7lh4\" (UniqueName: \"kubernetes.io/projected/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-kube-api-access-p7lh4\") pod \"7cbce4a4-b5ac-481c-87a7-47bc07e3361a\" (UID: \"7cbce4a4-b5ac-481c-87a7-47bc07e3361a\") " Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.581927 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5w9cx\" (UniqueName: \"kubernetes.io/projected/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-kube-api-access-5w9cx\") pod \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.581974 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-combined-ca-bundle\") pod \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.582002 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-scripts\") pod \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.582052 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-combined-ca-bundle\") pod \"7cbce4a4-b5ac-481c-87a7-47bc07e3361a\" (UID: \"7cbce4a4-b5ac-481c-87a7-47bc07e3361a\") " Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.582079 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-logs\") pod \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\" (UID: \"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58\") " Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.582821 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-logs" (OuterVolumeSpecName: "logs") pod "60ccdbf7-fc2d-4292-8ff6-d388ccda2b58" (UID: "60ccdbf7-fc2d-4292-8ff6-d388ccda2b58"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.587949 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-scripts" (OuterVolumeSpecName: "scripts") pod "60ccdbf7-fc2d-4292-8ff6-d388ccda2b58" (UID: "60ccdbf7-fc2d-4292-8ff6-d388ccda2b58"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.588038 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-kube-api-access-5w9cx" (OuterVolumeSpecName: "kube-api-access-5w9cx") pod "60ccdbf7-fc2d-4292-8ff6-d388ccda2b58" (UID: "60ccdbf7-fc2d-4292-8ff6-d388ccda2b58"). InnerVolumeSpecName "kube-api-access-5w9cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.588244 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-kube-api-access-p7lh4" (OuterVolumeSpecName: "kube-api-access-p7lh4") pod "7cbce4a4-b5ac-481c-87a7-47bc07e3361a" (UID: "7cbce4a4-b5ac-481c-87a7-47bc07e3361a"). InnerVolumeSpecName "kube-api-access-p7lh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.592304 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7cbce4a4-b5ac-481c-87a7-47bc07e3361a" (UID: "7cbce4a4-b5ac-481c-87a7-47bc07e3361a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.608566 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-config-data" (OuterVolumeSpecName: "config-data") pod "60ccdbf7-fc2d-4292-8ff6-d388ccda2b58" (UID: "60ccdbf7-fc2d-4292-8ff6-d388ccda2b58"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.613456 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7cbce4a4-b5ac-481c-87a7-47bc07e3361a" (UID: "7cbce4a4-b5ac-481c-87a7-47bc07e3361a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.621820 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60ccdbf7-fc2d-4292-8ff6-d388ccda2b58" (UID: "60ccdbf7-fc2d-4292-8ff6-d388ccda2b58"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.684320 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.684353 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.684365 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.684378 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.684388 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.684396 4745 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.684404 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7lh4\" (UniqueName: \"kubernetes.io/projected/7cbce4a4-b5ac-481c-87a7-47bc07e3361a-kube-api-access-p7lh4\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.684414 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5w9cx\" (UniqueName: \"kubernetes.io/projected/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58-kube-api-access-5w9cx\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.776871 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.865085 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-vhj7g"] Oct 10 13:34:18 crc kubenswrapper[4745]: I1010 13:34:18.865357 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" podUID="2abd8e35-90b0-49c6-935e-2b757228070b" containerName="dnsmasq-dns" containerID="cri-o://d4cc78df2a3fa8942a88a570f62105212fdec51468aac828e752d0db38ffdc8d" gracePeriod=10 Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.112305 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-c2vrp" event={"ID":"60ccdbf7-fc2d-4292-8ff6-d388ccda2b58","Type":"ContainerDied","Data":"4ad2f8ae54ad3d4d5468a74615131926a7ce5eaf5a65eb17bd8419fbd905baa0"} Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.112345 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ad2f8ae54ad3d4d5468a74615131926a7ce5eaf5a65eb17bd8419fbd905baa0" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.112397 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-c2vrp" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.114723 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dd6lk" event={"ID":"7cbce4a4-b5ac-481c-87a7-47bc07e3361a","Type":"ContainerDied","Data":"aed9f254148e3d7868baf87730bca7496d68d58a7b3dcfb65e39cbab3eecaa37"} Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.114761 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aed9f254148e3d7868baf87730bca7496d68d58a7b3dcfb65e39cbab3eecaa37" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.114793 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dd6lk" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.117397 4745 generic.go:334] "Generic (PLEG): container finished" podID="2abd8e35-90b0-49c6-935e-2b757228070b" containerID="d4cc78df2a3fa8942a88a570f62105212fdec51468aac828e752d0db38ffdc8d" exitCode=0 Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.117445 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" event={"ID":"2abd8e35-90b0-49c6-935e-2b757228070b","Type":"ContainerDied","Data":"d4cc78df2a3fa8942a88a570f62105212fdec51468aac828e752d0db38ffdc8d"} Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.576683 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7b8bc58bc6-p44zs"] Oct 10 13:34:19 crc kubenswrapper[4745]: E1010 13:34:19.579005 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cbce4a4-b5ac-481c-87a7-47bc07e3361a" containerName="barbican-db-sync" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.579024 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cbce4a4-b5ac-481c-87a7-47bc07e3361a" containerName="barbican-db-sync" Oct 10 13:34:19 crc kubenswrapper[4745]: E1010 13:34:19.579061 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83abb158-a66f-48ed-ae55-40a2c35cfbf2" containerName="horizon" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.579070 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="83abb158-a66f-48ed-ae55-40a2c35cfbf2" containerName="horizon" Oct 10 13:34:19 crc kubenswrapper[4745]: E1010 13:34:19.579089 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60ccdbf7-fc2d-4292-8ff6-d388ccda2b58" containerName="placement-db-sync" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.579098 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="60ccdbf7-fc2d-4292-8ff6-d388ccda2b58" containerName="placement-db-sync" Oct 10 13:34:19 crc kubenswrapper[4745]: E1010 13:34:19.579110 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17718d89-5706-452c-a482-027b2c6f7c0d" containerName="horizon" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.579119 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="17718d89-5706-452c-a482-027b2c6f7c0d" containerName="horizon" Oct 10 13:34:19 crc kubenswrapper[4745]: E1010 13:34:19.579134 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83abb158-a66f-48ed-ae55-40a2c35cfbf2" containerName="horizon-log" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.579143 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="83abb158-a66f-48ed-ae55-40a2c35cfbf2" containerName="horizon-log" Oct 10 13:34:19 crc kubenswrapper[4745]: E1010 13:34:19.579160 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c688f752-0686-412c-85f8-8304cc245608" containerName="horizon" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.579169 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="c688f752-0686-412c-85f8-8304cc245608" containerName="horizon" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.579385 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="c688f752-0686-412c-85f8-8304cc245608" containerName="horizon" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.579403 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="83abb158-a66f-48ed-ae55-40a2c35cfbf2" containerName="horizon" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.579421 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="17718d89-5706-452c-a482-027b2c6f7c0d" containerName="horizon" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.579439 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cbce4a4-b5ac-481c-87a7-47bc07e3361a" containerName="barbican-db-sync" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.579450 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="83abb158-a66f-48ed-ae55-40a2c35cfbf2" containerName="horizon-log" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.579465 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="60ccdbf7-fc2d-4292-8ff6-d388ccda2b58" containerName="placement-db-sync" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.580544 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.582381 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.582547 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-9dnj8" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.582899 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.583962 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.584786 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.596977 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7b8bc58bc6-p44zs"] Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.704548 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cf84964-c89c-4990-8320-823b10461186-public-tls-certs\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.704931 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cf84964-c89c-4990-8320-823b10461186-scripts\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.705050 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf84964-c89c-4990-8320-823b10461186-config-data\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.705171 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cf84964-c89c-4990-8320-823b10461186-internal-tls-certs\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.705282 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cf84964-c89c-4990-8320-823b10461186-combined-ca-bundle\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.705421 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cf84964-c89c-4990-8320-823b10461186-logs\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.705539 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nkp8\" (UniqueName: \"kubernetes.io/projected/5cf84964-c89c-4990-8320-823b10461186-kube-api-access-7nkp8\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.770601 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-57477cc59-r674m"] Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.772602 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-57477cc59-r674m" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.776118 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.776194 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.776353 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-hw7zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.793178 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-57477cc59-r674m"] Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.807366 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cf84964-c89c-4990-8320-823b10461186-public-tls-certs\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.807476 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cf84964-c89c-4990-8320-823b10461186-scripts\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.807516 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf84964-c89c-4990-8320-823b10461186-config-data\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.807561 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cf84964-c89c-4990-8320-823b10461186-internal-tls-certs\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.807592 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cf84964-c89c-4990-8320-823b10461186-combined-ca-bundle\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.807637 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cf84964-c89c-4990-8320-823b10461186-logs\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.807671 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nkp8\" (UniqueName: \"kubernetes.io/projected/5cf84964-c89c-4990-8320-823b10461186-kube-api-access-7nkp8\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.808466 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cf84964-c89c-4990-8320-823b10461186-logs\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.814139 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf84964-c89c-4990-8320-823b10461186-config-data\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.814982 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cf84964-c89c-4990-8320-823b10461186-internal-tls-certs\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.819717 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cf84964-c89c-4990-8320-823b10461186-scripts\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.824770 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-568f47fdfd-mnxcr"] Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.825182 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cf84964-c89c-4990-8320-823b10461186-combined-ca-bundle\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.826221 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cf84964-c89c-4990-8320-823b10461186-public-tls-certs\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.831956 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.838777 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.850448 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nkp8\" (UniqueName: \"kubernetes.io/projected/5cf84964-c89c-4990-8320-823b10461186-kube-api-access-7nkp8\") pod \"placement-7b8bc58bc6-p44zs\" (UID: \"5cf84964-c89c-4990-8320-823b10461186\") " pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.858608 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-568f47fdfd-mnxcr"] Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.909147 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-9845r"] Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.910054 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2195e7f7-9227-404a-8123-3d487c93ac6b-logs\") pod \"barbican-keystone-listener-568f47fdfd-mnxcr\" (UID: \"2195e7f7-9227-404a-8123-3d487c93ac6b\") " pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.910092 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k42fg\" (UniqueName: \"kubernetes.io/projected/2195e7f7-9227-404a-8123-3d487c93ac6b-kube-api-access-k42fg\") pod \"barbican-keystone-listener-568f47fdfd-mnxcr\" (UID: \"2195e7f7-9227-404a-8123-3d487c93ac6b\") " pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.910133 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82ee4ff2-4fa9-42fc-91dd-ee171c443ffa-logs\") pod \"barbican-worker-57477cc59-r674m\" (UID: \"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa\") " pod="openstack/barbican-worker-57477cc59-r674m" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.910164 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/82ee4ff2-4fa9-42fc-91dd-ee171c443ffa-config-data-custom\") pod \"barbican-worker-57477cc59-r674m\" (UID: \"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa\") " pod="openstack/barbican-worker-57477cc59-r674m" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.910187 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2195e7f7-9227-404a-8123-3d487c93ac6b-combined-ca-bundle\") pod \"barbican-keystone-listener-568f47fdfd-mnxcr\" (UID: \"2195e7f7-9227-404a-8123-3d487c93ac6b\") " pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.910209 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xgqq\" (UniqueName: \"kubernetes.io/projected/82ee4ff2-4fa9-42fc-91dd-ee171c443ffa-kube-api-access-2xgqq\") pod \"barbican-worker-57477cc59-r674m\" (UID: \"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa\") " pod="openstack/barbican-worker-57477cc59-r674m" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.910242 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ee4ff2-4fa9-42fc-91dd-ee171c443ffa-config-data\") pod \"barbican-worker-57477cc59-r674m\" (UID: \"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa\") " pod="openstack/barbican-worker-57477cc59-r674m" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.910263 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2195e7f7-9227-404a-8123-3d487c93ac6b-config-data\") pod \"barbican-keystone-listener-568f47fdfd-mnxcr\" (UID: \"2195e7f7-9227-404a-8123-3d487c93ac6b\") " pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.910296 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2195e7f7-9227-404a-8123-3d487c93ac6b-config-data-custom\") pod \"barbican-keystone-listener-568f47fdfd-mnxcr\" (UID: \"2195e7f7-9227-404a-8123-3d487c93ac6b\") " pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.910327 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ee4ff2-4fa9-42fc-91dd-ee171c443ffa-combined-ca-bundle\") pod \"barbican-worker-57477cc59-r674m\" (UID: \"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa\") " pod="openstack/barbican-worker-57477cc59-r674m" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.910597 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.922066 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-9845r"] Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.935288 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:19 crc kubenswrapper[4745]: I1010 13:34:19.999454 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-644bf58b5b-rhgpb"] Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.003705 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.013679 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.013748 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.013782 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ee4ff2-4fa9-42fc-91dd-ee171c443ffa-combined-ca-bundle\") pod \"barbican-worker-57477cc59-r674m\" (UID: \"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa\") " pod="openstack/barbican-worker-57477cc59-r674m" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.013852 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-config\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.013894 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2195e7f7-9227-404a-8123-3d487c93ac6b-logs\") pod \"barbican-keystone-listener-568f47fdfd-mnxcr\" (UID: \"2195e7f7-9227-404a-8123-3d487c93ac6b\") " pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.013911 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k42fg\" (UniqueName: \"kubernetes.io/projected/2195e7f7-9227-404a-8123-3d487c93ac6b-kube-api-access-k42fg\") pod \"barbican-keystone-listener-568f47fdfd-mnxcr\" (UID: \"2195e7f7-9227-404a-8123-3d487c93ac6b\") " pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.013948 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.013972 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82ee4ff2-4fa9-42fc-91dd-ee171c443ffa-logs\") pod \"barbican-worker-57477cc59-r674m\" (UID: \"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa\") " pod="openstack/barbican-worker-57477cc59-r674m" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.014004 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ssz8\" (UniqueName: \"kubernetes.io/projected/aedee3a1-1870-4691-a5a4-d5321b90e3a2-kube-api-access-8ssz8\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.014033 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/82ee4ff2-4fa9-42fc-91dd-ee171c443ffa-config-data-custom\") pod \"barbican-worker-57477cc59-r674m\" (UID: \"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa\") " pod="openstack/barbican-worker-57477cc59-r674m" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.014059 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2195e7f7-9227-404a-8123-3d487c93ac6b-combined-ca-bundle\") pod \"barbican-keystone-listener-568f47fdfd-mnxcr\" (UID: \"2195e7f7-9227-404a-8123-3d487c93ac6b\") " pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.014092 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xgqq\" (UniqueName: \"kubernetes.io/projected/82ee4ff2-4fa9-42fc-91dd-ee171c443ffa-kube-api-access-2xgqq\") pod \"barbican-worker-57477cc59-r674m\" (UID: \"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa\") " pod="openstack/barbican-worker-57477cc59-r674m" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.014125 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.014159 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ee4ff2-4fa9-42fc-91dd-ee171c443ffa-config-data\") pod \"barbican-worker-57477cc59-r674m\" (UID: \"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa\") " pod="openstack/barbican-worker-57477cc59-r674m" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.014186 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2195e7f7-9227-404a-8123-3d487c93ac6b-config-data\") pod \"barbican-keystone-listener-568f47fdfd-mnxcr\" (UID: \"2195e7f7-9227-404a-8123-3d487c93ac6b\") " pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.014236 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2195e7f7-9227-404a-8123-3d487c93ac6b-config-data-custom\") pod \"barbican-keystone-listener-568f47fdfd-mnxcr\" (UID: \"2195e7f7-9227-404a-8123-3d487c93ac6b\") " pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.015482 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82ee4ff2-4fa9-42fc-91dd-ee171c443ffa-logs\") pod \"barbican-worker-57477cc59-r674m\" (UID: \"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa\") " pod="openstack/barbican-worker-57477cc59-r674m" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.016623 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2195e7f7-9227-404a-8123-3d487c93ac6b-logs\") pod \"barbican-keystone-listener-568f47fdfd-mnxcr\" (UID: \"2195e7f7-9227-404a-8123-3d487c93ac6b\") " pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.019881 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.042136 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2195e7f7-9227-404a-8123-3d487c93ac6b-combined-ca-bundle\") pod \"barbican-keystone-listener-568f47fdfd-mnxcr\" (UID: \"2195e7f7-9227-404a-8123-3d487c93ac6b\") " pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.043716 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xgqq\" (UniqueName: \"kubernetes.io/projected/82ee4ff2-4fa9-42fc-91dd-ee171c443ffa-kube-api-access-2xgqq\") pod \"barbican-worker-57477cc59-r674m\" (UID: \"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa\") " pod="openstack/barbican-worker-57477cc59-r674m" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.045369 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ee4ff2-4fa9-42fc-91dd-ee171c443ffa-combined-ca-bundle\") pod \"barbican-worker-57477cc59-r674m\" (UID: \"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa\") " pod="openstack/barbican-worker-57477cc59-r674m" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.045958 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2195e7f7-9227-404a-8123-3d487c93ac6b-config-data-custom\") pod \"barbican-keystone-listener-568f47fdfd-mnxcr\" (UID: \"2195e7f7-9227-404a-8123-3d487c93ac6b\") " pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.047208 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k42fg\" (UniqueName: \"kubernetes.io/projected/2195e7f7-9227-404a-8123-3d487c93ac6b-kube-api-access-k42fg\") pod \"barbican-keystone-listener-568f47fdfd-mnxcr\" (UID: \"2195e7f7-9227-404a-8123-3d487c93ac6b\") " pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.050543 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/82ee4ff2-4fa9-42fc-91dd-ee171c443ffa-config-data-custom\") pod \"barbican-worker-57477cc59-r674m\" (UID: \"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa\") " pod="openstack/barbican-worker-57477cc59-r674m" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.061638 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ee4ff2-4fa9-42fc-91dd-ee171c443ffa-config-data\") pod \"barbican-worker-57477cc59-r674m\" (UID: \"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa\") " pod="openstack/barbican-worker-57477cc59-r674m" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.064739 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-644bf58b5b-rhgpb"] Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.067854 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2195e7f7-9227-404a-8123-3d487c93ac6b-config-data\") pod \"barbican-keystone-listener-568f47fdfd-mnxcr\" (UID: \"2195e7f7-9227-404a-8123-3d487c93ac6b\") " pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.093489 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-57477cc59-r674m" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.115882 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.116246 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ssz8\" (UniqueName: \"kubernetes.io/projected/aedee3a1-1870-4691-a5a4-d5321b90e3a2-kube-api-access-8ssz8\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.116349 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-combined-ca-bundle\") pod \"barbican-api-644bf58b5b-rhgpb\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.116516 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.116629 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-config-data-custom\") pod \"barbican-api-644bf58b5b-rhgpb\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.116901 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.116998 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-config-data\") pod \"barbican-api-644bf58b5b-rhgpb\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.117181 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.122077 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-config\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.122202 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce71cf3a-de78-452f-b48d-fd8edd70ef35-logs\") pod \"barbican-api-644bf58b5b-rhgpb\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.122298 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft74c\" (UniqueName: \"kubernetes.io/projected/ce71cf3a-de78-452f-b48d-fd8edd70ef35-kube-api-access-ft74c\") pod \"barbican-api-644bf58b5b-rhgpb\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.117939 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.119005 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.118511 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.120597 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.123988 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-config\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.136529 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ssz8\" (UniqueName: \"kubernetes.io/projected/aedee3a1-1870-4691-a5a4-d5321b90e3a2-kube-api-access-8ssz8\") pod \"dnsmasq-dns-75c8ddd69c-9845r\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.224368 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-config-data-custom\") pod \"barbican-api-644bf58b5b-rhgpb\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.224454 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-config-data\") pod \"barbican-api-644bf58b5b-rhgpb\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.224497 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce71cf3a-de78-452f-b48d-fd8edd70ef35-logs\") pod \"barbican-api-644bf58b5b-rhgpb\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.224520 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft74c\" (UniqueName: \"kubernetes.io/projected/ce71cf3a-de78-452f-b48d-fd8edd70ef35-kube-api-access-ft74c\") pod \"barbican-api-644bf58b5b-rhgpb\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.224569 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-combined-ca-bundle\") pod \"barbican-api-644bf58b5b-rhgpb\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.225569 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce71cf3a-de78-452f-b48d-fd8edd70ef35-logs\") pod \"barbican-api-644bf58b5b-rhgpb\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.229852 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-config-data-custom\") pod \"barbican-api-644bf58b5b-rhgpb\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.230626 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-config-data\") pod \"barbican-api-644bf58b5b-rhgpb\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.231087 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-combined-ca-bundle\") pod \"barbican-api-644bf58b5b-rhgpb\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.242439 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft74c\" (UniqueName: \"kubernetes.io/projected/ce71cf3a-de78-452f-b48d-fd8edd70ef35-kube-api-access-ft74c\") pod \"barbican-api-644bf58b5b-rhgpb\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.249748 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.263049 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:20 crc kubenswrapper[4745]: I1010 13:34:20.430633 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:21 crc kubenswrapper[4745]: I1010 13:34:21.880572 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:34:21 crc kubenswrapper[4745]: I1010 13:34:21.954957 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-config\") pod \"2abd8e35-90b0-49c6-935e-2b757228070b\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " Oct 10 13:34:21 crc kubenswrapper[4745]: I1010 13:34:21.955001 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-ovsdbserver-nb\") pod \"2abd8e35-90b0-49c6-935e-2b757228070b\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " Oct 10 13:34:21 crc kubenswrapper[4745]: I1010 13:34:21.955057 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-dns-svc\") pod \"2abd8e35-90b0-49c6-935e-2b757228070b\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " Oct 10 13:34:21 crc kubenswrapper[4745]: I1010 13:34:21.955080 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-ovsdbserver-sb\") pod \"2abd8e35-90b0-49c6-935e-2b757228070b\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " Oct 10 13:34:21 crc kubenswrapper[4745]: I1010 13:34:21.955255 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-dns-swift-storage-0\") pod \"2abd8e35-90b0-49c6-935e-2b757228070b\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " Oct 10 13:34:21 crc kubenswrapper[4745]: I1010 13:34:21.955301 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f658\" (UniqueName: \"kubernetes.io/projected/2abd8e35-90b0-49c6-935e-2b757228070b-kube-api-access-9f658\") pod \"2abd8e35-90b0-49c6-935e-2b757228070b\" (UID: \"2abd8e35-90b0-49c6-935e-2b757228070b\") " Oct 10 13:34:21 crc kubenswrapper[4745]: I1010 13:34:21.977217 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2abd8e35-90b0-49c6-935e-2b757228070b-kube-api-access-9f658" (OuterVolumeSpecName: "kube-api-access-9f658") pod "2abd8e35-90b0-49c6-935e-2b757228070b" (UID: "2abd8e35-90b0-49c6-935e-2b757228070b"). InnerVolumeSpecName "kube-api-access-9f658". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.036781 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2abd8e35-90b0-49c6-935e-2b757228070b" (UID: "2abd8e35-90b0-49c6-935e-2b757228070b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.047443 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2abd8e35-90b0-49c6-935e-2b757228070b" (UID: "2abd8e35-90b0-49c6-935e-2b757228070b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.051407 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-config" (OuterVolumeSpecName: "config") pod "2abd8e35-90b0-49c6-935e-2b757228070b" (UID: "2abd8e35-90b0-49c6-935e-2b757228070b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.056796 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f658\" (UniqueName: \"kubernetes.io/projected/2abd8e35-90b0-49c6-935e-2b757228070b-kube-api-access-9f658\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.056826 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.056837 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.056846 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.076232 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2abd8e35-90b0-49c6-935e-2b757228070b" (UID: "2abd8e35-90b0-49c6-935e-2b757228070b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.085790 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2abd8e35-90b0-49c6-935e-2b757228070b" (UID: "2abd8e35-90b0-49c6-935e-2b757228070b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.151000 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" event={"ID":"2abd8e35-90b0-49c6-935e-2b757228070b","Type":"ContainerDied","Data":"897fc06e54b380cf61aae7fe4f88536659f49f95c27e8cbad74e4d5b65925224"} Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.151075 4745 scope.go:117] "RemoveContainer" containerID="d4cc78df2a3fa8942a88a570f62105212fdec51468aac828e752d0db38ffdc8d" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.151290 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-vhj7g" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.162810 4745 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.162833 4745 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2abd8e35-90b0-49c6-935e-2b757228070b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.235142 4745 scope.go:117] "RemoveContainer" containerID="fe6923548dc550c4b9297c117d5d4c0d356b9280c25223c732f7c5d716df1791" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.243050 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-vhj7g"] Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.251275 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-vhj7g"] Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.365303 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-644bf58b5b-rhgpb"] Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.400175 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-795bd49d76-9gqws"] Oct 10 13:34:22 crc kubenswrapper[4745]: E1010 13:34:22.400795 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2abd8e35-90b0-49c6-935e-2b757228070b" containerName="dnsmasq-dns" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.400815 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="2abd8e35-90b0-49c6-935e-2b757228070b" containerName="dnsmasq-dns" Oct 10 13:34:22 crc kubenswrapper[4745]: E1010 13:34:22.400851 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2abd8e35-90b0-49c6-935e-2b757228070b" containerName="init" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.400876 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="2abd8e35-90b0-49c6-935e-2b757228070b" containerName="init" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.401142 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="2abd8e35-90b0-49c6-935e-2b757228070b" containerName="dnsmasq-dns" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.402398 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.406398 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.406632 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.409470 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-795bd49d76-9gqws"] Oct 10 13:34:22 crc kubenswrapper[4745]: E1010 13:34:22.412932 4745 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45009f2e_d0d5_4a20_b62c_d5d0534e1551.slice/crio-a85b3684c58b6faeef3d56c54eb1502b33619172c3103963b89eca08d70b6f4c\": RecentStats: unable to find data in memory cache]" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.477671 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b21c656-1b66-422f-8397-12c630ee4889-logs\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.477714 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b21c656-1b66-422f-8397-12c630ee4889-config-data\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.477757 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b21c656-1b66-422f-8397-12c630ee4889-public-tls-certs\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.477782 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b21c656-1b66-422f-8397-12c630ee4889-internal-tls-certs\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.477804 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b21c656-1b66-422f-8397-12c630ee4889-combined-ca-bundle\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.477840 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b21c656-1b66-422f-8397-12c630ee4889-config-data-custom\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.477881 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97xdn\" (UniqueName: \"kubernetes.io/projected/5b21c656-1b66-422f-8397-12c630ee4889-kube-api-access-97xdn\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.583367 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b21c656-1b66-422f-8397-12c630ee4889-logs\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.583664 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b21c656-1b66-422f-8397-12c630ee4889-config-data\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.583683 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b21c656-1b66-422f-8397-12c630ee4889-public-tls-certs\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.583700 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b21c656-1b66-422f-8397-12c630ee4889-internal-tls-certs\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.583719 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b21c656-1b66-422f-8397-12c630ee4889-combined-ca-bundle\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.583776 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b21c656-1b66-422f-8397-12c630ee4889-config-data-custom\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.583819 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97xdn\" (UniqueName: \"kubernetes.io/projected/5b21c656-1b66-422f-8397-12c630ee4889-kube-api-access-97xdn\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.583873 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b21c656-1b66-422f-8397-12c630ee4889-logs\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.588610 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b21c656-1b66-422f-8397-12c630ee4889-combined-ca-bundle\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.588989 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b21c656-1b66-422f-8397-12c630ee4889-public-tls-certs\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.589439 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b21c656-1b66-422f-8397-12c630ee4889-config-data\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.591707 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b21c656-1b66-422f-8397-12c630ee4889-config-data-custom\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.593852 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b21c656-1b66-422f-8397-12c630ee4889-internal-tls-certs\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.610408 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97xdn\" (UniqueName: \"kubernetes.io/projected/5b21c656-1b66-422f-8397-12c630ee4889-kube-api-access-97xdn\") pod \"barbican-api-795bd49d76-9gqws\" (UID: \"5b21c656-1b66-422f-8397-12c630ee4889\") " pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: W1010 13:34:22.667015 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82ee4ff2_4fa9_42fc_91dd_ee171c443ffa.slice/crio-5b61d3c45a944587799eecc0f854bf94956901aa0a29fecd42aca157b769b941 WatchSource:0}: Error finding container 5b61d3c45a944587799eecc0f854bf94956901aa0a29fecd42aca157b769b941: Status 404 returned error can't find the container with id 5b61d3c45a944587799eecc0f854bf94956901aa0a29fecd42aca157b769b941 Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.674786 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7b8bc58bc6-p44zs"] Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.690519 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-568f47fdfd-mnxcr"] Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.698071 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-57477cc59-r674m"] Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.733781 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.778652 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2abd8e35-90b0-49c6-935e-2b757228070b" path="/var/lib/kubelet/pods/2abd8e35-90b0-49c6-935e-2b757228070b/volumes" Oct 10 13:34:22 crc kubenswrapper[4745]: I1010 13:34:22.788076 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-9845r"] Oct 10 13:34:22 crc kubenswrapper[4745]: W1010 13:34:22.790865 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaedee3a1_1870_4691_a5a4_d5321b90e3a2.slice/crio-11514f5707a97f5582307623c99788163cc04927c98b858d6ce8c54cda6e99e1 WatchSource:0}: Error finding container 11514f5707a97f5582307623c99788163cc04927c98b858d6ce8c54cda6e99e1: Status 404 returned error can't find the container with id 11514f5707a97f5582307623c99788163cc04927c98b858d6ce8c54cda6e99e1 Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.168254 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"822d9514-5b19-440a-b4dc-5850f2bc8549","Type":"ContainerStarted","Data":"98a8cd228708b5bff3188b921a97a008342706a73879272cf1dc7f4bb73d2873"} Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.169649 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerName="proxy-httpd" containerID="cri-o://98a8cd228708b5bff3188b921a97a008342706a73879272cf1dc7f4bb73d2873" gracePeriod=30 Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.169642 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerName="ceilometer-central-agent" containerID="cri-o://fb5d0bae521c81290ed4fc0fa35e521dba6cb81fc58a9fd2a68445497fdd0c2d" gracePeriod=30 Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.169781 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerName="sg-core" containerID="cri-o://4695a67f02f9e358f724e4f6181357e4d97926fa6e352f01577f0d646306bddc" gracePeriod=30 Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.169830 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerName="ceilometer-notification-agent" containerID="cri-o://1772984a3fb04be75513af6fa054a11290ec24ca5f0b9560a2313d44f5b05cdd" gracePeriod=30 Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.176165 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b8bc58bc6-p44zs" event={"ID":"5cf84964-c89c-4990-8320-823b10461186","Type":"ContainerStarted","Data":"bf34b4b50b6bdf100e031c39871b5f6f329ce045904d2ff90778ced62564153e"} Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.176280 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b8bc58bc6-p44zs" event={"ID":"5cf84964-c89c-4990-8320-823b10461186","Type":"ContainerStarted","Data":"6d7c9cad6f4d43700670687f09bc026909a65f6115ebdda48040ca4596a57dac"} Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.194902 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.019417051 podStartE2EDuration="1m2.194886084s" podCreationTimestamp="2025-10-10 13:33:21 +0000 UTC" firstStartedPulling="2025-10-10 13:33:22.752597241 +0000 UTC m=+916.650254004" lastFinishedPulling="2025-10-10 13:34:21.928066274 +0000 UTC m=+975.825723037" observedRunningTime="2025-10-10 13:34:23.194434133 +0000 UTC m=+977.092090896" watchObservedRunningTime="2025-10-10 13:34:23.194886084 +0000 UTC m=+977.092542837" Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.198113 4745 generic.go:334] "Generic (PLEG): container finished" podID="aedee3a1-1870-4691-a5a4-d5321b90e3a2" containerID="2106c08fdc3dac8310716e5f79bc6e7c90283f3b8f4c24b3383765acc57f7386" exitCode=0 Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.198259 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" event={"ID":"aedee3a1-1870-4691-a5a4-d5321b90e3a2","Type":"ContainerDied","Data":"2106c08fdc3dac8310716e5f79bc6e7c90283f3b8f4c24b3383765acc57f7386"} Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.198357 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" event={"ID":"aedee3a1-1870-4691-a5a4-d5321b90e3a2","Type":"ContainerStarted","Data":"11514f5707a97f5582307623c99788163cc04927c98b858d6ce8c54cda6e99e1"} Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.206057 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57477cc59-r674m" event={"ID":"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa","Type":"ContainerStarted","Data":"5b61d3c45a944587799eecc0f854bf94956901aa0a29fecd42aca157b769b941"} Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.207675 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" event={"ID":"2195e7f7-9227-404a-8123-3d487c93ac6b","Type":"ContainerStarted","Data":"1a17afaae4b2b9c5e23e8a23690a02c3e49805b157404cfcaaa7439f4f2db044"} Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.209336 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-644bf58b5b-rhgpb" event={"ID":"ce71cf3a-de78-452f-b48d-fd8edd70ef35","Type":"ContainerStarted","Data":"2167d93fb5d197090a7afb89b1f6c920a5dcfb7790b293b2d176ca0b2f2e1ad0"} Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.209354 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-644bf58b5b-rhgpb" event={"ID":"ce71cf3a-de78-452f-b48d-fd8edd70ef35","Type":"ContainerStarted","Data":"b102f6fab43c0340292ed8434bb7be605f71b341ed968bde26964ddfd2344a59"} Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.209366 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-644bf58b5b-rhgpb" event={"ID":"ce71cf3a-de78-452f-b48d-fd8edd70ef35","Type":"ContainerStarted","Data":"3063c8dfc65caaa2abcc092ac785436ff0dfec11b71c6f30cbab7fc0e33b9655"} Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.209971 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.210112 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.269950 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-644bf58b5b-rhgpb" podStartSLOduration=4.26992702 podStartE2EDuration="4.26992702s" podCreationTimestamp="2025-10-10 13:34:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:34:23.269193131 +0000 UTC m=+977.166849894" watchObservedRunningTime="2025-10-10 13:34:23.26992702 +0000 UTC m=+977.167583773" Oct 10 13:34:23 crc kubenswrapper[4745]: I1010 13:34:23.277878 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-795bd49d76-9gqws"] Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.104545 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-745bfbd6-ms6d2" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.233180 4745 generic.go:334] "Generic (PLEG): container finished" podID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerID="98a8cd228708b5bff3188b921a97a008342706a73879272cf1dc7f4bb73d2873" exitCode=0 Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.233209 4745 generic.go:334] "Generic (PLEG): container finished" podID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerID="4695a67f02f9e358f724e4f6181357e4d97926fa6e352f01577f0d646306bddc" exitCode=2 Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.233217 4745 generic.go:334] "Generic (PLEG): container finished" podID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerID="fb5d0bae521c81290ed4fc0fa35e521dba6cb81fc58a9fd2a68445497fdd0c2d" exitCode=0 Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.233263 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"822d9514-5b19-440a-b4dc-5850f2bc8549","Type":"ContainerDied","Data":"98a8cd228708b5bff3188b921a97a008342706a73879272cf1dc7f4bb73d2873"} Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.233289 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"822d9514-5b19-440a-b4dc-5850f2bc8549","Type":"ContainerDied","Data":"4695a67f02f9e358f724e4f6181357e4d97926fa6e352f01577f0d646306bddc"} Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.233299 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"822d9514-5b19-440a-b4dc-5850f2bc8549","Type":"ContainerDied","Data":"fb5d0bae521c81290ed4fc0fa35e521dba6cb81fc58a9fd2a68445497fdd0c2d"} Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.238039 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b8bc58bc6-p44zs" event={"ID":"5cf84964-c89c-4990-8320-823b10461186","Type":"ContainerStarted","Data":"a7cf9628488168429508ce09f253c60071142466e97995abbb2e4147b037b050"} Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.239180 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.239224 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.256374 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" event={"ID":"aedee3a1-1870-4691-a5a4-d5321b90e3a2","Type":"ContainerStarted","Data":"6fb283d89359918d31200b911555af2620c5f6b69137d2e59804ed5768af383f"} Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.257407 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.276176 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7b8bc58bc6-p44zs" podStartSLOduration=5.275551222 podStartE2EDuration="5.275551222s" podCreationTimestamp="2025-10-10 13:34:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:34:24.273999973 +0000 UTC m=+978.171656736" watchObservedRunningTime="2025-10-10 13:34:24.275551222 +0000 UTC m=+978.173207985" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.292772 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795bd49d76-9gqws" event={"ID":"5b21c656-1b66-422f-8397-12c630ee4889","Type":"ContainerStarted","Data":"11fc76a2ed15beeb471d538c83ee6f4b0df5ed10206d19bb8daaba748bdbeca7"} Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.292814 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795bd49d76-9gqws" event={"ID":"5b21c656-1b66-422f-8397-12c630ee4889","Type":"ContainerStarted","Data":"bb4c90248b24b944ca864485f08975312116ed9dc0f8a16c35d249e79b39741e"} Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.292825 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795bd49d76-9gqws" event={"ID":"5b21c656-1b66-422f-8397-12c630ee4889","Type":"ContainerStarted","Data":"f3f1db6b4f9815f9022956c13fbbbf4f9d825befccb6e00a8c358274e7b51bd3"} Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.294562 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.294587 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.315037 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" podStartSLOduration=5.315018908 podStartE2EDuration="5.315018908s" podCreationTimestamp="2025-10-10 13:34:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:34:24.297409698 +0000 UTC m=+978.195066461" watchObservedRunningTime="2025-10-10 13:34:24.315018908 +0000 UTC m=+978.212675671" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.322086 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-795bd49d76-9gqws" podStartSLOduration=2.322069804 podStartE2EDuration="2.322069804s" podCreationTimestamp="2025-10-10 13:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:34:24.316165857 +0000 UTC m=+978.213822620" watchObservedRunningTime="2025-10-10 13:34:24.322069804 +0000 UTC m=+978.219726567" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.330295 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vskzt" event={"ID":"7e86aa13-6741-4a98-88f7-1fec0870468f","Type":"ContainerStarted","Data":"94cec6aba343eb4bb33675324772ba9f235d6b31718fcf8704c4fd2f5e178063"} Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.354054 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-vskzt" podStartSLOduration=2.931495109 podStartE2EDuration="39.354038183s" podCreationTimestamp="2025-10-10 13:33:45 +0000 UTC" firstStartedPulling="2025-10-10 13:33:46.80633969 +0000 UTC m=+940.703996453" lastFinishedPulling="2025-10-10 13:34:23.228882764 +0000 UTC m=+977.126539527" observedRunningTime="2025-10-10 13:34:24.350157106 +0000 UTC m=+978.247813859" watchObservedRunningTime="2025-10-10 13:34:24.354038183 +0000 UTC m=+978.251694946" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.771326 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.772478 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.778135 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.778358 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-th4lv" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.778632 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.782010 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.867325 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7gng\" (UniqueName: \"kubernetes.io/projected/5cc91ec2-6a94-4923-82cf-d488c077a97b-kube-api-access-h7gng\") pod \"openstackclient\" (UID: \"5cc91ec2-6a94-4923-82cf-d488c077a97b\") " pod="openstack/openstackclient" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.867490 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5cc91ec2-6a94-4923-82cf-d488c077a97b-openstack-config-secret\") pod \"openstackclient\" (UID: \"5cc91ec2-6a94-4923-82cf-d488c077a97b\") " pod="openstack/openstackclient" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.867952 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5cc91ec2-6a94-4923-82cf-d488c077a97b-openstack-config\") pod \"openstackclient\" (UID: \"5cc91ec2-6a94-4923-82cf-d488c077a97b\") " pod="openstack/openstackclient" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.868040 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cc91ec2-6a94-4923-82cf-d488c077a97b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5cc91ec2-6a94-4923-82cf-d488c077a97b\") " pod="openstack/openstackclient" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.970624 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5cc91ec2-6a94-4923-82cf-d488c077a97b-openstack-config\") pod \"openstackclient\" (UID: \"5cc91ec2-6a94-4923-82cf-d488c077a97b\") " pod="openstack/openstackclient" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.971017 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cc91ec2-6a94-4923-82cf-d488c077a97b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5cc91ec2-6a94-4923-82cf-d488c077a97b\") " pod="openstack/openstackclient" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.972063 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5cc91ec2-6a94-4923-82cf-d488c077a97b-openstack-config\") pod \"openstackclient\" (UID: \"5cc91ec2-6a94-4923-82cf-d488c077a97b\") " pod="openstack/openstackclient" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.972558 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7gng\" (UniqueName: \"kubernetes.io/projected/5cc91ec2-6a94-4923-82cf-d488c077a97b-kube-api-access-h7gng\") pod \"openstackclient\" (UID: \"5cc91ec2-6a94-4923-82cf-d488c077a97b\") " pod="openstack/openstackclient" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.972757 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5cc91ec2-6a94-4923-82cf-d488c077a97b-openstack-config-secret\") pod \"openstackclient\" (UID: \"5cc91ec2-6a94-4923-82cf-d488c077a97b\") " pod="openstack/openstackclient" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.980430 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cc91ec2-6a94-4923-82cf-d488c077a97b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5cc91ec2-6a94-4923-82cf-d488c077a97b\") " pod="openstack/openstackclient" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.980641 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5cc91ec2-6a94-4923-82cf-d488c077a97b-openstack-config-secret\") pod \"openstackclient\" (UID: \"5cc91ec2-6a94-4923-82cf-d488c077a97b\") " pod="openstack/openstackclient" Oct 10 13:34:24 crc kubenswrapper[4745]: I1010 13:34:24.993497 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7gng\" (UniqueName: \"kubernetes.io/projected/5cc91ec2-6a94-4923-82cf-d488c077a97b-kube-api-access-h7gng\") pod \"openstackclient\" (UID: \"5cc91ec2-6a94-4923-82cf-d488c077a97b\") " pod="openstack/openstackclient" Oct 10 13:34:25 crc kubenswrapper[4745]: I1010 13:34:25.241500 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 13:34:25 crc kubenswrapper[4745]: I1010 13:34:25.338852 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57477cc59-r674m" event={"ID":"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa","Type":"ContainerStarted","Data":"d823c3dfb477fa134011e0b209dec6ede34131a98894004a77e49b4d1f7358f1"} Oct 10 13:34:25 crc kubenswrapper[4745]: I1010 13:34:25.341120 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" event={"ID":"2195e7f7-9227-404a-8123-3d487c93ac6b","Type":"ContainerStarted","Data":"023f6f5af0015e493c973e43fd9122e54cddeadf0cd56c318990f557fb24d85f"} Oct 10 13:34:25 crc kubenswrapper[4745]: I1010 13:34:25.341186 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" event={"ID":"2195e7f7-9227-404a-8123-3d487c93ac6b","Type":"ContainerStarted","Data":"352e7860a1babbc49dec6512ad5588437fc1aaea0a5ece20ea906259c8bd2d61"} Oct 10 13:34:25 crc kubenswrapper[4745]: I1010 13:34:25.370752 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-568f47fdfd-mnxcr" podStartSLOduration=4.393509997 podStartE2EDuration="6.370722001s" podCreationTimestamp="2025-10-10 13:34:19 +0000 UTC" firstStartedPulling="2025-10-10 13:34:22.678035067 +0000 UTC m=+976.575691830" lastFinishedPulling="2025-10-10 13:34:24.655247071 +0000 UTC m=+978.552903834" observedRunningTime="2025-10-10 13:34:25.362653219 +0000 UTC m=+979.260309972" watchObservedRunningTime="2025-10-10 13:34:25.370722001 +0000 UTC m=+979.268378764" Oct 10 13:34:25 crc kubenswrapper[4745]: I1010 13:34:25.768116 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 10 13:34:26 crc kubenswrapper[4745]: I1010 13:34:26.351381 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"5cc91ec2-6a94-4923-82cf-d488c077a97b","Type":"ContainerStarted","Data":"d637e005158339d8f5dd93643173b8b3e325f538282b01c71a0fb75296c860e2"} Oct 10 13:34:26 crc kubenswrapper[4745]: I1010 13:34:26.353546 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57477cc59-r674m" event={"ID":"82ee4ff2-4fa9-42fc-91dd-ee171c443ffa","Type":"ContainerStarted","Data":"84fe7a8ea6beedb450a0bac6d8cb6f9e4318f9f48ce006174fb2afdba312ea3a"} Oct 10 13:34:26 crc kubenswrapper[4745]: I1010 13:34:26.371981 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-57477cc59-r674m" podStartSLOduration=5.39354894 podStartE2EDuration="7.371961394s" podCreationTimestamp="2025-10-10 13:34:19 +0000 UTC" firstStartedPulling="2025-10-10 13:34:22.673111454 +0000 UTC m=+976.570768217" lastFinishedPulling="2025-10-10 13:34:24.651523908 +0000 UTC m=+978.549180671" observedRunningTime="2025-10-10 13:34:26.36861749 +0000 UTC m=+980.266274253" watchObservedRunningTime="2025-10-10 13:34:26.371961394 +0000 UTC m=+980.269618157" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.043968 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.128294 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-combined-ca-bundle\") pod \"822d9514-5b19-440a-b4dc-5850f2bc8549\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.128362 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-scripts\") pod \"822d9514-5b19-440a-b4dc-5850f2bc8549\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.128416 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/822d9514-5b19-440a-b4dc-5850f2bc8549-log-httpd\") pod \"822d9514-5b19-440a-b4dc-5850f2bc8549\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.128454 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-config-data\") pod \"822d9514-5b19-440a-b4dc-5850f2bc8549\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.128487 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wz9r\" (UniqueName: \"kubernetes.io/projected/822d9514-5b19-440a-b4dc-5850f2bc8549-kube-api-access-2wz9r\") pod \"822d9514-5b19-440a-b4dc-5850f2bc8549\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.128554 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/822d9514-5b19-440a-b4dc-5850f2bc8549-run-httpd\") pod \"822d9514-5b19-440a-b4dc-5850f2bc8549\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.128582 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-sg-core-conf-yaml\") pod \"822d9514-5b19-440a-b4dc-5850f2bc8549\" (UID: \"822d9514-5b19-440a-b4dc-5850f2bc8549\") " Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.129191 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/822d9514-5b19-440a-b4dc-5850f2bc8549-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "822d9514-5b19-440a-b4dc-5850f2bc8549" (UID: "822d9514-5b19-440a-b4dc-5850f2bc8549"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.129300 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/822d9514-5b19-440a-b4dc-5850f2bc8549-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "822d9514-5b19-440a-b4dc-5850f2bc8549" (UID: "822d9514-5b19-440a-b4dc-5850f2bc8549"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.137398 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/822d9514-5b19-440a-b4dc-5850f2bc8549-kube-api-access-2wz9r" (OuterVolumeSpecName: "kube-api-access-2wz9r") pod "822d9514-5b19-440a-b4dc-5850f2bc8549" (UID: "822d9514-5b19-440a-b4dc-5850f2bc8549"). InnerVolumeSpecName "kube-api-access-2wz9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.141012 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-scripts" (OuterVolumeSpecName: "scripts") pod "822d9514-5b19-440a-b4dc-5850f2bc8549" (UID: "822d9514-5b19-440a-b4dc-5850f2bc8549"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.160962 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "822d9514-5b19-440a-b4dc-5850f2bc8549" (UID: "822d9514-5b19-440a-b4dc-5850f2bc8549"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.213692 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "822d9514-5b19-440a-b4dc-5850f2bc8549" (UID: "822d9514-5b19-440a-b4dc-5850f2bc8549"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.218430 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-config-data" (OuterVolumeSpecName: "config-data") pod "822d9514-5b19-440a-b4dc-5850f2bc8549" (UID: "822d9514-5b19-440a-b4dc-5850f2bc8549"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.230448 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.230486 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.230497 4745 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/822d9514-5b19-440a-b4dc-5850f2bc8549-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.230507 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.230516 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wz9r\" (UniqueName: \"kubernetes.io/projected/822d9514-5b19-440a-b4dc-5850f2bc8549-kube-api-access-2wz9r\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.230528 4745 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/822d9514-5b19-440a-b4dc-5850f2bc8549-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.230536 4745 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/822d9514-5b19-440a-b4dc-5850f2bc8549-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.375946 4745 generic.go:334] "Generic (PLEG): container finished" podID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerID="1772984a3fb04be75513af6fa054a11290ec24ca5f0b9560a2313d44f5b05cdd" exitCode=0 Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.376019 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"822d9514-5b19-440a-b4dc-5850f2bc8549","Type":"ContainerDied","Data":"1772984a3fb04be75513af6fa054a11290ec24ca5f0b9560a2313d44f5b05cdd"} Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.376049 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.376068 4745 scope.go:117] "RemoveContainer" containerID="98a8cd228708b5bff3188b921a97a008342706a73879272cf1dc7f4bb73d2873" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.376056 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"822d9514-5b19-440a-b4dc-5850f2bc8549","Type":"ContainerDied","Data":"ae3bdb198f4bac50a10007e2b392a3a634e62adcd0eb931cf2cd264d3b9cb7e2"} Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.380950 4745 generic.go:334] "Generic (PLEG): container finished" podID="7e86aa13-6741-4a98-88f7-1fec0870468f" containerID="94cec6aba343eb4bb33675324772ba9f235d6b31718fcf8704c4fd2f5e178063" exitCode=0 Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.380977 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vskzt" event={"ID":"7e86aa13-6741-4a98-88f7-1fec0870468f","Type":"ContainerDied","Data":"94cec6aba343eb4bb33675324772ba9f235d6b31718fcf8704c4fd2f5e178063"} Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.410642 4745 scope.go:117] "RemoveContainer" containerID="4695a67f02f9e358f724e4f6181357e4d97926fa6e352f01577f0d646306bddc" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.430609 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.438713 4745 scope.go:117] "RemoveContainer" containerID="1772984a3fb04be75513af6fa054a11290ec24ca5f0b9560a2313d44f5b05cdd" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.441928 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.449367 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:34:28 crc kubenswrapper[4745]: E1010 13:34:28.449769 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerName="ceilometer-notification-agent" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.449786 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerName="ceilometer-notification-agent" Oct 10 13:34:28 crc kubenswrapper[4745]: E1010 13:34:28.449807 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerName="ceilometer-central-agent" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.449813 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerName="ceilometer-central-agent" Oct 10 13:34:28 crc kubenswrapper[4745]: E1010 13:34:28.449830 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerName="proxy-httpd" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.449836 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerName="proxy-httpd" Oct 10 13:34:28 crc kubenswrapper[4745]: E1010 13:34:28.449852 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerName="sg-core" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.449858 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerName="sg-core" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.450009 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerName="proxy-httpd" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.450027 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerName="sg-core" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.450037 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerName="ceilometer-central-agent" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.450048 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="822d9514-5b19-440a-b4dc-5850f2bc8549" containerName="ceilometer-notification-agent" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.451614 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.453630 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.455528 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.458949 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.480207 4745 scope.go:117] "RemoveContainer" containerID="fb5d0bae521c81290ed4fc0fa35e521dba6cb81fc58a9fd2a68445497fdd0c2d" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.512608 4745 scope.go:117] "RemoveContainer" containerID="98a8cd228708b5bff3188b921a97a008342706a73879272cf1dc7f4bb73d2873" Oct 10 13:34:28 crc kubenswrapper[4745]: E1010 13:34:28.514453 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98a8cd228708b5bff3188b921a97a008342706a73879272cf1dc7f4bb73d2873\": container with ID starting with 98a8cd228708b5bff3188b921a97a008342706a73879272cf1dc7f4bb73d2873 not found: ID does not exist" containerID="98a8cd228708b5bff3188b921a97a008342706a73879272cf1dc7f4bb73d2873" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.514495 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98a8cd228708b5bff3188b921a97a008342706a73879272cf1dc7f4bb73d2873"} err="failed to get container status \"98a8cd228708b5bff3188b921a97a008342706a73879272cf1dc7f4bb73d2873\": rpc error: code = NotFound desc = could not find container \"98a8cd228708b5bff3188b921a97a008342706a73879272cf1dc7f4bb73d2873\": container with ID starting with 98a8cd228708b5bff3188b921a97a008342706a73879272cf1dc7f4bb73d2873 not found: ID does not exist" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.514516 4745 scope.go:117] "RemoveContainer" containerID="4695a67f02f9e358f724e4f6181357e4d97926fa6e352f01577f0d646306bddc" Oct 10 13:34:28 crc kubenswrapper[4745]: E1010 13:34:28.515193 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4695a67f02f9e358f724e4f6181357e4d97926fa6e352f01577f0d646306bddc\": container with ID starting with 4695a67f02f9e358f724e4f6181357e4d97926fa6e352f01577f0d646306bddc not found: ID does not exist" containerID="4695a67f02f9e358f724e4f6181357e4d97926fa6e352f01577f0d646306bddc" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.515214 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4695a67f02f9e358f724e4f6181357e4d97926fa6e352f01577f0d646306bddc"} err="failed to get container status \"4695a67f02f9e358f724e4f6181357e4d97926fa6e352f01577f0d646306bddc\": rpc error: code = NotFound desc = could not find container \"4695a67f02f9e358f724e4f6181357e4d97926fa6e352f01577f0d646306bddc\": container with ID starting with 4695a67f02f9e358f724e4f6181357e4d97926fa6e352f01577f0d646306bddc not found: ID does not exist" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.515226 4745 scope.go:117] "RemoveContainer" containerID="1772984a3fb04be75513af6fa054a11290ec24ca5f0b9560a2313d44f5b05cdd" Oct 10 13:34:28 crc kubenswrapper[4745]: E1010 13:34:28.515615 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1772984a3fb04be75513af6fa054a11290ec24ca5f0b9560a2313d44f5b05cdd\": container with ID starting with 1772984a3fb04be75513af6fa054a11290ec24ca5f0b9560a2313d44f5b05cdd not found: ID does not exist" containerID="1772984a3fb04be75513af6fa054a11290ec24ca5f0b9560a2313d44f5b05cdd" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.515650 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1772984a3fb04be75513af6fa054a11290ec24ca5f0b9560a2313d44f5b05cdd"} err="failed to get container status \"1772984a3fb04be75513af6fa054a11290ec24ca5f0b9560a2313d44f5b05cdd\": rpc error: code = NotFound desc = could not find container \"1772984a3fb04be75513af6fa054a11290ec24ca5f0b9560a2313d44f5b05cdd\": container with ID starting with 1772984a3fb04be75513af6fa054a11290ec24ca5f0b9560a2313d44f5b05cdd not found: ID does not exist" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.515668 4745 scope.go:117] "RemoveContainer" containerID="fb5d0bae521c81290ed4fc0fa35e521dba6cb81fc58a9fd2a68445497fdd0c2d" Oct 10 13:34:28 crc kubenswrapper[4745]: E1010 13:34:28.516103 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb5d0bae521c81290ed4fc0fa35e521dba6cb81fc58a9fd2a68445497fdd0c2d\": container with ID starting with fb5d0bae521c81290ed4fc0fa35e521dba6cb81fc58a9fd2a68445497fdd0c2d not found: ID does not exist" containerID="fb5d0bae521c81290ed4fc0fa35e521dba6cb81fc58a9fd2a68445497fdd0c2d" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.516162 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb5d0bae521c81290ed4fc0fa35e521dba6cb81fc58a9fd2a68445497fdd0c2d"} err="failed to get container status \"fb5d0bae521c81290ed4fc0fa35e521dba6cb81fc58a9fd2a68445497fdd0c2d\": rpc error: code = NotFound desc = could not find container \"fb5d0bae521c81290ed4fc0fa35e521dba6cb81fc58a9fd2a68445497fdd0c2d\": container with ID starting with fb5d0bae521c81290ed4fc0fa35e521dba6cb81fc58a9fd2a68445497fdd0c2d not found: ID does not exist" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.539965 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-run-httpd\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.540053 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.540230 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.540444 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-log-httpd\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.540511 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-scripts\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.540564 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-config-data\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.540612 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lphmj\" (UniqueName: \"kubernetes.io/projected/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-kube-api-access-lphmj\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.642506 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.642628 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-log-httpd\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.642676 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-scripts\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.642777 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-config-data\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.643691 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lphmj\" (UniqueName: \"kubernetes.io/projected/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-kube-api-access-lphmj\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.643702 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-log-httpd\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.643833 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-run-httpd\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.643900 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.644094 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-run-httpd\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.648590 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.653206 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-scripts\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.661753 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-config-data\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.664482 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lphmj\" (UniqueName: \"kubernetes.io/projected/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-kube-api-access-lphmj\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.666894 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " pod="openstack/ceilometer-0" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.757828 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="822d9514-5b19-440a-b4dc-5850f2bc8549" path="/var/lib/kubelet/pods/822d9514-5b19-440a-b4dc-5850f2bc8549/volumes" Oct 10 13:34:28 crc kubenswrapper[4745]: I1010 13:34:28.780144 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.302868 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:34:29 crc kubenswrapper[4745]: W1010 13:34:29.345321 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf664e2bc_73b1_464c_acb7_e5ad00a75ec7.slice/crio-b598dbed87005c8bcfe5ba58a857b6243a866d05b2914fdd7ab1b2e0ea2e1948 WatchSource:0}: Error finding container b598dbed87005c8bcfe5ba58a857b6243a866d05b2914fdd7ab1b2e0ea2e1948: Status 404 returned error can't find the container with id b598dbed87005c8bcfe5ba58a857b6243a866d05b2914fdd7ab1b2e0ea2e1948 Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.388818 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5d77644dd7-jv6wz"] Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.390597 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.394888 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.395102 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.401831 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.403562 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5d77644dd7-jv6wz"] Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.406524 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f664e2bc-73b1-464c-acb7-e5ad00a75ec7","Type":"ContainerStarted","Data":"b598dbed87005c8bcfe5ba58a857b6243a866d05b2914fdd7ab1b2e0ea2e1948"} Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.491908 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2vw4\" (UniqueName: \"kubernetes.io/projected/6aa7ade8-3d4b-4168-80a9-673c73b919a3-kube-api-access-n2vw4\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.492264 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6aa7ade8-3d4b-4168-80a9-673c73b919a3-run-httpd\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.492299 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa7ade8-3d4b-4168-80a9-673c73b919a3-internal-tls-certs\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.492332 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6aa7ade8-3d4b-4168-80a9-673c73b919a3-etc-swift\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.492350 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aa7ade8-3d4b-4168-80a9-673c73b919a3-config-data\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.492371 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa7ade8-3d4b-4168-80a9-673c73b919a3-public-tls-certs\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.492398 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6aa7ade8-3d4b-4168-80a9-673c73b919a3-log-httpd\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.492424 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa7ade8-3d4b-4168-80a9-673c73b919a3-combined-ca-bundle\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.598367 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6aa7ade8-3d4b-4168-80a9-673c73b919a3-etc-swift\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.598653 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aa7ade8-3d4b-4168-80a9-673c73b919a3-config-data\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.598693 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa7ade8-3d4b-4168-80a9-673c73b919a3-public-tls-certs\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.598756 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6aa7ade8-3d4b-4168-80a9-673c73b919a3-log-httpd\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.598814 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa7ade8-3d4b-4168-80a9-673c73b919a3-combined-ca-bundle\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.599575 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2vw4\" (UniqueName: \"kubernetes.io/projected/6aa7ade8-3d4b-4168-80a9-673c73b919a3-kube-api-access-n2vw4\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.599679 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6aa7ade8-3d4b-4168-80a9-673c73b919a3-run-httpd\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.599720 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa7ade8-3d4b-4168-80a9-673c73b919a3-internal-tls-certs\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.600213 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6aa7ade8-3d4b-4168-80a9-673c73b919a3-run-httpd\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.600300 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6aa7ade8-3d4b-4168-80a9-673c73b919a3-log-httpd\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.610700 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aa7ade8-3d4b-4168-80a9-673c73b919a3-config-data\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.612720 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa7ade8-3d4b-4168-80a9-673c73b919a3-combined-ca-bundle\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.613410 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa7ade8-3d4b-4168-80a9-673c73b919a3-internal-tls-certs\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.613675 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6aa7ade8-3d4b-4168-80a9-673c73b919a3-etc-swift\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.617502 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa7ade8-3d4b-4168-80a9-673c73b919a3-public-tls-certs\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.627313 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2vw4\" (UniqueName: \"kubernetes.io/projected/6aa7ade8-3d4b-4168-80a9-673c73b919a3-kube-api-access-n2vw4\") pod \"swift-proxy-5d77644dd7-jv6wz\" (UID: \"6aa7ade8-3d4b-4168-80a9-673c73b919a3\") " pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.736666 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:29 crc kubenswrapper[4745]: I1010 13:34:29.887904 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vskzt" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.011190 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e86aa13-6741-4a98-88f7-1fec0870468f-etc-machine-id\") pod \"7e86aa13-6741-4a98-88f7-1fec0870468f\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.011308 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-scripts\") pod \"7e86aa13-6741-4a98-88f7-1fec0870468f\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.011375 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ttxx\" (UniqueName: \"kubernetes.io/projected/7e86aa13-6741-4a98-88f7-1fec0870468f-kube-api-access-9ttxx\") pod \"7e86aa13-6741-4a98-88f7-1fec0870468f\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.011399 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-config-data\") pod \"7e86aa13-6741-4a98-88f7-1fec0870468f\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.011497 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-combined-ca-bundle\") pod \"7e86aa13-6741-4a98-88f7-1fec0870468f\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.011523 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-db-sync-config-data\") pod \"7e86aa13-6741-4a98-88f7-1fec0870468f\" (UID: \"7e86aa13-6741-4a98-88f7-1fec0870468f\") " Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.018354 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7e86aa13-6741-4a98-88f7-1fec0870468f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7e86aa13-6741-4a98-88f7-1fec0870468f" (UID: "7e86aa13-6741-4a98-88f7-1fec0870468f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.018621 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7e86aa13-6741-4a98-88f7-1fec0870468f" (UID: "7e86aa13-6741-4a98-88f7-1fec0870468f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.021948 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e86aa13-6741-4a98-88f7-1fec0870468f-kube-api-access-9ttxx" (OuterVolumeSpecName: "kube-api-access-9ttxx") pod "7e86aa13-6741-4a98-88f7-1fec0870468f" (UID: "7e86aa13-6741-4a98-88f7-1fec0870468f"). InnerVolumeSpecName "kube-api-access-9ttxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.028479 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-scripts" (OuterVolumeSpecName: "scripts") pod "7e86aa13-6741-4a98-88f7-1fec0870468f" (UID: "7e86aa13-6741-4a98-88f7-1fec0870468f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.058470 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e86aa13-6741-4a98-88f7-1fec0870468f" (UID: "7e86aa13-6741-4a98-88f7-1fec0870468f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.113272 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.113296 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ttxx\" (UniqueName: \"kubernetes.io/projected/7e86aa13-6741-4a98-88f7-1fec0870468f-kube-api-access-9ttxx\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.113305 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.113313 4745 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.113323 4745 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e86aa13-6741-4a98-88f7-1fec0870468f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.130625 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-z2v8n"] Oct 10 13:34:30 crc kubenswrapper[4745]: E1010 13:34:30.131117 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e86aa13-6741-4a98-88f7-1fec0870468f" containerName="cinder-db-sync" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.131136 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e86aa13-6741-4a98-88f7-1fec0870468f" containerName="cinder-db-sync" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.131331 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e86aa13-6741-4a98-88f7-1fec0870468f" containerName="cinder-db-sync" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.132086 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-z2v8n" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.133718 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-config-data" (OuterVolumeSpecName: "config-data") pod "7e86aa13-6741-4a98-88f7-1fec0870468f" (UID: "7e86aa13-6741-4a98-88f7-1fec0870468f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.144974 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-z2v8n"] Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.214346 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dtnk\" (UniqueName: \"kubernetes.io/projected/cc3cdd18-7362-4610-8f9b-7444bd45b09c-kube-api-access-8dtnk\") pod \"nova-api-db-create-z2v8n\" (UID: \"cc3cdd18-7362-4610-8f9b-7444bd45b09c\") " pod="openstack/nova-api-db-create-z2v8n" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.214432 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e86aa13-6741-4a98-88f7-1fec0870468f-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.226258 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-xc9ls"] Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.227414 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xc9ls" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.238769 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-xc9ls"] Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.269942 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.347705 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dtnk\" (UniqueName: \"kubernetes.io/projected/cc3cdd18-7362-4610-8f9b-7444bd45b09c-kube-api-access-8dtnk\") pod \"nova-api-db-create-z2v8n\" (UID: \"cc3cdd18-7362-4610-8f9b-7444bd45b09c\") " pod="openstack/nova-api-db-create-z2v8n" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.347870 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld6qx\" (UniqueName: \"kubernetes.io/projected/28904cbe-09fb-499a-8d88-23ac50034c90-kube-api-access-ld6qx\") pod \"nova-cell0-db-create-xc9ls\" (UID: \"28904cbe-09fb-499a-8d88-23ac50034c90\") " pod="openstack/nova-cell0-db-create-xc9ls" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.421503 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dtnk\" (UniqueName: \"kubernetes.io/projected/cc3cdd18-7362-4610-8f9b-7444bd45b09c-kube-api-access-8dtnk\") pod \"nova-api-db-create-z2v8n\" (UID: \"cc3cdd18-7362-4610-8f9b-7444bd45b09c\") " pod="openstack/nova-api-db-create-z2v8n" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.424227 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-762hz"] Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.424803 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b966f6c9-762hz" podUID="55673aa8-01ac-4a68-b059-cf1a1c0579d6" containerName="dnsmasq-dns" containerID="cri-o://e745532a6a236899e940920e329204fed2f608738e54734da0e920f0a4765c9e" gracePeriod=10 Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.432053 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vskzt" event={"ID":"7e86aa13-6741-4a98-88f7-1fec0870468f","Type":"ContainerDied","Data":"c36b6d79fedb4cc7d59fd78cd5dee16082b1ffb26e970e4595bb6dd1f7d9b814"} Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.432166 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vskzt" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.432186 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c36b6d79fedb4cc7d59fd78cd5dee16082b1ffb26e970e4595bb6dd1f7d9b814" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.442183 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f664e2bc-73b1-464c-acb7-e5ad00a75ec7","Type":"ContainerStarted","Data":"78e1401b81bb10d24e922e63ca98ca3b70cdb6c3f17f89b5d58cfd6cbf914358"} Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.444699 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5d77644dd7-jv6wz" event={"ID":"6aa7ade8-3d4b-4168-80a9-673c73b919a3","Type":"ContainerStarted","Data":"2e7f790ef1c92dd3c5151b6b641b7308e71c3d67d594a7cbc3e834335249939c"} Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.447341 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5d77644dd7-jv6wz"] Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.450122 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld6qx\" (UniqueName: \"kubernetes.io/projected/28904cbe-09fb-499a-8d88-23ac50034c90-kube-api-access-ld6qx\") pod \"nova-cell0-db-create-xc9ls\" (UID: \"28904cbe-09fb-499a-8d88-23ac50034c90\") " pod="openstack/nova-cell0-db-create-xc9ls" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.458262 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-z2v8n" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.461205 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-h8ngs"] Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.462429 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-h8ngs" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.464993 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld6qx\" (UniqueName: \"kubernetes.io/projected/28904cbe-09fb-499a-8d88-23ac50034c90-kube-api-access-ld6qx\") pod \"nova-cell0-db-create-xc9ls\" (UID: \"28904cbe-09fb-499a-8d88-23ac50034c90\") " pod="openstack/nova-cell0-db-create-xc9ls" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.487296 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-h8ngs"] Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.562074 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgzf5\" (UniqueName: \"kubernetes.io/projected/a3771609-ba86-4413-8cbd-3d266cd0380d-kube-api-access-qgzf5\") pod \"nova-cell1-db-create-h8ngs\" (UID: \"a3771609-ba86-4413-8cbd-3d266cd0380d\") " pod="openstack/nova-cell1-db-create-h8ngs" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.569508 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.577773 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xc9ls" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.663746 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgzf5\" (UniqueName: \"kubernetes.io/projected/a3771609-ba86-4413-8cbd-3d266cd0380d-kube-api-access-qgzf5\") pod \"nova-cell1-db-create-h8ngs\" (UID: \"a3771609-ba86-4413-8cbd-3d266cd0380d\") " pod="openstack/nova-cell1-db-create-h8ngs" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.679704 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.681256 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.685010 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.685542 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wjzhr" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.687275 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.687500 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.696675 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgzf5\" (UniqueName: \"kubernetes.io/projected/a3771609-ba86-4413-8cbd-3d266cd0380d-kube-api-access-qgzf5\") pod \"nova-cell1-db-create-h8ngs\" (UID: \"a3771609-ba86-4413-8cbd-3d266cd0380d\") " pod="openstack/nova-cell1-db-create-h8ngs" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.703465 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.814643 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-rshbk"] Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.816081 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-rshbk"] Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.816161 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.869100 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-h8ngs" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.869585 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-config-data\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.869629 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.869693 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nl4v\" (UniqueName: \"kubernetes.io/projected/7560bd3f-0029-4878-bedf-a47020316cca-kube-api-access-7nl4v\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.869714 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-scripts\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.869744 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.869760 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7560bd3f-0029-4878-bedf-a47020316cca-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.944065 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.945684 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.951068 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.968404 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.971591 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-config\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.971633 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.971670 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-config-data\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.971712 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.971770 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5h4c\" (UniqueName: \"kubernetes.io/projected/16739af7-770b-4253-be87-ca62b9c21fb9-kube-api-access-v5h4c\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.971792 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-dns-svc\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.971838 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nl4v\" (UniqueName: \"kubernetes.io/projected/7560bd3f-0029-4878-bedf-a47020316cca-kube-api-access-7nl4v\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.971857 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.971876 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-scripts\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.971892 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.971908 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7560bd3f-0029-4878-bedf-a47020316cca-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.971935 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.974375 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7560bd3f-0029-4878-bedf-a47020316cca-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.976827 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-config-data\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.978540 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.988153 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-scripts\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:30 crc kubenswrapper[4745]: I1010 13:34:30.999535 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nl4v\" (UniqueName: \"kubernetes.io/projected/7560bd3f-0029-4878-bedf-a47020316cca-kube-api-access-7nl4v\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.010028 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.073943 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n92qt\" (UniqueName: \"kubernetes.io/projected/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-kube-api-access-n92qt\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.073976 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-logs\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.074003 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-config\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.074036 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.074056 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-scripts\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.074088 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-config-data\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.074107 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5h4c\" (UniqueName: \"kubernetes.io/projected/16739af7-770b-4253-be87-ca62b9c21fb9-kube-api-access-v5h4c\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.074123 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.074141 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-dns-svc\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.074160 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.074208 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.074235 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-config-data-custom\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.074260 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.075045 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.075318 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-config\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.075858 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.086390 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-dns-svc\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.092675 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.095050 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5h4c\" (UniqueName: \"kubernetes.io/projected/16739af7-770b-4253-be87-ca62b9c21fb9-kube-api-access-v5h4c\") pod \"dnsmasq-dns-5784cf869f-rshbk\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.145414 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.172585 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.176151 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-config-data-custom\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.176303 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n92qt\" (UniqueName: \"kubernetes.io/projected/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-kube-api-access-n92qt\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.176344 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-logs\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.176408 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-scripts\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.176474 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-config-data\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.176513 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.176567 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.176721 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.183110 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-logs\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.184343 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-config-data\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.186405 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-config-data-custom\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.187679 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.191153 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-z2v8n"] Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.195170 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-scripts\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.208170 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n92qt\" (UniqueName: \"kubernetes.io/projected/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-kube-api-access-n92qt\") pod \"cinder-api-0\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.471162 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.492912 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.499858 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f664e2bc-73b1-464c-acb7-e5ad00a75ec7","Type":"ContainerStarted","Data":"20b58afd107e76758d855767ab2ab220abf53dc8a5540faee28ac0d54f08b0ed"} Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.516889 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5d77644dd7-jv6wz" event={"ID":"6aa7ade8-3d4b-4168-80a9-673c73b919a3","Type":"ContainerStarted","Data":"e2353d824a151722b01e7c6629f1be81734b6324884d9786106d0f466b90dfc9"} Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.516924 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5d77644dd7-jv6wz" event={"ID":"6aa7ade8-3d4b-4168-80a9-673c73b919a3","Type":"ContainerStarted","Data":"ebe62ba5bf262b146488233c520c9261387d60ead1ff89c8c4603c4c9b3670b3"} Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.517639 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.517660 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.536943 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-z2v8n" event={"ID":"cc3cdd18-7362-4610-8f9b-7444bd45b09c","Type":"ContainerStarted","Data":"3a9373993164362ccbd907a0a0fbf672b8e36f79d061a855e5b5bc7cf434cb2b"} Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.592395 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-528v8\" (UniqueName: \"kubernetes.io/projected/55673aa8-01ac-4a68-b059-cf1a1c0579d6-kube-api-access-528v8\") pod \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.592515 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-config\") pod \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.592564 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-ovsdbserver-nb\") pod \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.592603 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-ovsdbserver-sb\") pod \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.592762 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-dns-svc\") pod \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.592791 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-dns-swift-storage-0\") pod \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\" (UID: \"55673aa8-01ac-4a68-b059-cf1a1c0579d6\") " Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.607003 4745 generic.go:334] "Generic (PLEG): container finished" podID="55673aa8-01ac-4a68-b059-cf1a1c0579d6" containerID="e745532a6a236899e940920e329204fed2f608738e54734da0e920f0a4765c9e" exitCode=0 Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.607052 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-762hz" event={"ID":"55673aa8-01ac-4a68-b059-cf1a1c0579d6","Type":"ContainerDied","Data":"e745532a6a236899e940920e329204fed2f608738e54734da0e920f0a4765c9e"} Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.607079 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-762hz" event={"ID":"55673aa8-01ac-4a68-b059-cf1a1c0579d6","Type":"ContainerDied","Data":"576a1186a5f0b7e385429ac91d4ed7d006654896d3f44fdb7c265538a2c25809"} Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.607095 4745 scope.go:117] "RemoveContainer" containerID="e745532a6a236899e940920e329204fed2f608738e54734da0e920f0a4765c9e" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.607252 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-762hz" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.646975 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55673aa8-01ac-4a68-b059-cf1a1c0579d6-kube-api-access-528v8" (OuterVolumeSpecName: "kube-api-access-528v8") pod "55673aa8-01ac-4a68-b059-cf1a1c0579d6" (UID: "55673aa8-01ac-4a68-b059-cf1a1c0579d6"). InnerVolumeSpecName "kube-api-access-528v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.709056 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-528v8\" (UniqueName: \"kubernetes.io/projected/55673aa8-01ac-4a68-b059-cf1a1c0579d6-kube-api-access-528v8\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.726058 4745 scope.go:117] "RemoveContainer" containerID="9f5fb23de37bb84f17c9eac9238bce93713ed1ab457df225d4dffde8aef9657c" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.728317 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5d77644dd7-jv6wz" podStartSLOduration=2.728306237 podStartE2EDuration="2.728306237s" podCreationTimestamp="2025-10-10 13:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:34:31.558299428 +0000 UTC m=+985.455956201" watchObservedRunningTime="2025-10-10 13:34:31.728306237 +0000 UTC m=+985.625963000" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.734441 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-h8ngs"] Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.744902 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-xc9ls"] Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.853079 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "55673aa8-01ac-4a68-b059-cf1a1c0579d6" (UID: "55673aa8-01ac-4a68-b059-cf1a1c0579d6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.918715 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.931124 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "55673aa8-01ac-4a68-b059-cf1a1c0579d6" (UID: "55673aa8-01ac-4a68-b059-cf1a1c0579d6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.939279 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "55673aa8-01ac-4a68-b059-cf1a1c0579d6" (UID: "55673aa8-01ac-4a68-b059-cf1a1c0579d6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.941465 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-config" (OuterVolumeSpecName: "config") pod "55673aa8-01ac-4a68-b059-cf1a1c0579d6" (UID: "55673aa8-01ac-4a68-b059-cf1a1c0579d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.953148 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "55673aa8-01ac-4a68-b059-cf1a1c0579d6" (UID: "55673aa8-01ac-4a68-b059-cf1a1c0579d6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:31 crc kubenswrapper[4745]: I1010 13:34:31.960675 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 13:34:32 crc kubenswrapper[4745]: I1010 13:34:32.020157 4745 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:32 crc kubenswrapper[4745]: I1010 13:34:32.020194 4745 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:32 crc kubenswrapper[4745]: I1010 13:34:32.020207 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:32 crc kubenswrapper[4745]: I1010 13:34:32.020217 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55673aa8-01ac-4a68-b059-cf1a1c0579d6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:32 crc kubenswrapper[4745]: I1010 13:34:32.217792 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-rshbk"] Oct 10 13:34:32 crc kubenswrapper[4745]: I1010 13:34:32.271390 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 10 13:34:32 crc kubenswrapper[4745]: I1010 13:34:32.331005 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-762hz"] Oct 10 13:34:32 crc kubenswrapper[4745]: I1010 13:34:32.361739 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-762hz"] Oct 10 13:34:32 crc kubenswrapper[4745]: I1010 13:34:32.665130 4745 generic.go:334] "Generic (PLEG): container finished" podID="cc3cdd18-7362-4610-8f9b-7444bd45b09c" containerID="f1531086e8b678df707da20bde70fd192a60eb1893dd4bc552ae427f3a109748" exitCode=0 Oct 10 13:34:32 crc kubenswrapper[4745]: I1010 13:34:32.665220 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-z2v8n" event={"ID":"cc3cdd18-7362-4610-8f9b-7444bd45b09c","Type":"ContainerDied","Data":"f1531086e8b678df707da20bde70fd192a60eb1893dd4bc552ae427f3a109748"} Oct 10 13:34:32 crc kubenswrapper[4745]: E1010 13:34:32.742061 4745 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45009f2e_d0d5_4a20_b62c_d5d0534e1551.slice/crio-a85b3684c58b6faeef3d56c54eb1502b33619172c3103963b89eca08d70b6f4c\": RecentStats: unable to find data in memory cache]" Oct 10 13:34:32 crc kubenswrapper[4745]: I1010 13:34:32.756861 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55673aa8-01ac-4a68-b059-cf1a1c0579d6" path="/var/lib/kubelet/pods/55673aa8-01ac-4a68-b059-cf1a1c0579d6/volumes" Oct 10 13:34:33 crc kubenswrapper[4745]: I1010 13:34:33.605130 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:33 crc kubenswrapper[4745]: I1010 13:34:33.689779 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 10 13:34:34 crc kubenswrapper[4745]: I1010 13:34:34.042038 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:34 crc kubenswrapper[4745]: I1010 13:34:34.718113 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:34 crc kubenswrapper[4745]: I1010 13:34:34.985295 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-795bd49d76-9gqws" Oct 10 13:34:35 crc kubenswrapper[4745]: I1010 13:34:35.027022 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-644bf58b5b-rhgpb"] Oct 10 13:34:35 crc kubenswrapper[4745]: I1010 13:34:35.027245 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-644bf58b5b-rhgpb" podUID="ce71cf3a-de78-452f-b48d-fd8edd70ef35" containerName="barbican-api-log" containerID="cri-o://b102f6fab43c0340292ed8434bb7be605f71b341ed968bde26964ddfd2344a59" gracePeriod=30 Oct 10 13:34:35 crc kubenswrapper[4745]: I1010 13:34:35.027343 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-644bf58b5b-rhgpb" podUID="ce71cf3a-de78-452f-b48d-fd8edd70ef35" containerName="barbican-api" containerID="cri-o://2167d93fb5d197090a7afb89b1f6c920a5dcfb7790b293b2d176ca0b2f2e1ad0" gracePeriod=30 Oct 10 13:34:35 crc kubenswrapper[4745]: I1010 13:34:35.040311 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-644bf58b5b-rhgpb" podUID="ce71cf3a-de78-452f-b48d-fd8edd70ef35" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": EOF" Oct 10 13:34:35 crc kubenswrapper[4745]: I1010 13:34:35.040347 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-644bf58b5b-rhgpb" podUID="ce71cf3a-de78-452f-b48d-fd8edd70ef35" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": EOF" Oct 10 13:34:35 crc kubenswrapper[4745]: I1010 13:34:35.726960 4745 generic.go:334] "Generic (PLEG): container finished" podID="ce71cf3a-de78-452f-b48d-fd8edd70ef35" containerID="b102f6fab43c0340292ed8434bb7be605f71b341ed968bde26964ddfd2344a59" exitCode=143 Oct 10 13:34:35 crc kubenswrapper[4745]: I1010 13:34:35.727039 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-644bf58b5b-rhgpb" event={"ID":"ce71cf3a-de78-452f-b48d-fd8edd70ef35","Type":"ContainerDied","Data":"b102f6fab43c0340292ed8434bb7be605f71b341ed968bde26964ddfd2344a59"} Oct 10 13:34:37 crc kubenswrapper[4745]: I1010 13:34:37.747043 4745 generic.go:334] "Generic (PLEG): container finished" podID="604fc859-90dd-4c8b-a216-6009b3dbcbf2" containerID="fec5f51d0b50e49ba472342af7f5e04b56f06dcd7a1d7afbd9ec19f6c386fb77" exitCode=137 Oct 10 13:34:37 crc kubenswrapper[4745]: I1010 13:34:37.747101 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65c7d6f5bb-9g4xm" event={"ID":"604fc859-90dd-4c8b-a216-6009b3dbcbf2","Type":"ContainerDied","Data":"fec5f51d0b50e49ba472342af7f5e04b56f06dcd7a1d7afbd9ec19f6c386fb77"} Oct 10 13:34:38 crc kubenswrapper[4745]: I1010 13:34:38.954276 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.447552 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-644bf58b5b-rhgpb" podUID="ce71cf3a-de78-452f-b48d-fd8edd70ef35" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": read tcp 10.217.0.2:49150->10.217.0.161:9311: read: connection reset by peer" Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.448244 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-644bf58b5b-rhgpb" podUID="ce71cf3a-de78-452f-b48d-fd8edd70ef35" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": read tcp 10.217.0.2:49156->10.217.0.161:9311: read: connection reset by peer" Oct 10 13:34:39 crc kubenswrapper[4745]: W1010 13:34:39.539956 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28904cbe_09fb_499a_8d88_23ac50034c90.slice/crio-0ae0f2edceb9a91db913ef974d3c44954a48e8aff621a844c5b6c995b581b738 WatchSource:0}: Error finding container 0ae0f2edceb9a91db913ef974d3c44954a48e8aff621a844c5b6c995b581b738: Status 404 returned error can't find the container with id 0ae0f2edceb9a91db913ef974d3c44954a48e8aff621a844c5b6c995b581b738 Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.581252 4745 scope.go:117] "RemoveContainer" containerID="e745532a6a236899e940920e329204fed2f608738e54734da0e920f0a4765c9e" Oct 10 13:34:39 crc kubenswrapper[4745]: E1010 13:34:39.586921 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e745532a6a236899e940920e329204fed2f608738e54734da0e920f0a4765c9e\": container with ID starting with e745532a6a236899e940920e329204fed2f608738e54734da0e920f0a4765c9e not found: ID does not exist" containerID="e745532a6a236899e940920e329204fed2f608738e54734da0e920f0a4765c9e" Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.586955 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e745532a6a236899e940920e329204fed2f608738e54734da0e920f0a4765c9e"} err="failed to get container status \"e745532a6a236899e940920e329204fed2f608738e54734da0e920f0a4765c9e\": rpc error: code = NotFound desc = could not find container \"e745532a6a236899e940920e329204fed2f608738e54734da0e920f0a4765c9e\": container with ID starting with e745532a6a236899e940920e329204fed2f608738e54734da0e920f0a4765c9e not found: ID does not exist" Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.586980 4745 scope.go:117] "RemoveContainer" containerID="9f5fb23de37bb84f17c9eac9238bce93713ed1ab457df225d4dffde8aef9657c" Oct 10 13:34:39 crc kubenswrapper[4745]: E1010 13:34:39.588179 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f5fb23de37bb84f17c9eac9238bce93713ed1ab457df225d4dffde8aef9657c\": container with ID starting with 9f5fb23de37bb84f17c9eac9238bce93713ed1ab457df225d4dffde8aef9657c not found: ID does not exist" containerID="9f5fb23de37bb84f17c9eac9238bce93713ed1ab457df225d4dffde8aef9657c" Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.588229 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f5fb23de37bb84f17c9eac9238bce93713ed1ab457df225d4dffde8aef9657c"} err="failed to get container status \"9f5fb23de37bb84f17c9eac9238bce93713ed1ab457df225d4dffde8aef9657c\": rpc error: code = NotFound desc = could not find container \"9f5fb23de37bb84f17c9eac9238bce93713ed1ab457df225d4dffde8aef9657c\": container with ID starting with 9f5fb23de37bb84f17c9eac9238bce93713ed1ab457df225d4dffde8aef9657c not found: ID does not exist" Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.806008 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-z2v8n" Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.806403 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7560bd3f-0029-4878-bedf-a47020316cca","Type":"ContainerStarted","Data":"3d6f561ffeb18c437cfef483d4d72dd510b62a3cbecfe016b0beaf750dc043e2"} Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.816175 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c","Type":"ContainerStarted","Data":"8133ca96d60e4913892109c6d38254b2f4668e24f05bd0d9ef4172bffe93fbbe"} Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.825265 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-h8ngs" event={"ID":"a3771609-ba86-4413-8cbd-3d266cd0380d","Type":"ContainerStarted","Data":"4e35f2b2847271b5a77b76833982a0adb24abd370f3311bc1c1483cd7b497716"} Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.827657 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xc9ls" event={"ID":"28904cbe-09fb-499a-8d88-23ac50034c90","Type":"ContainerStarted","Data":"0ae0f2edceb9a91db913ef974d3c44954a48e8aff621a844c5b6c995b581b738"} Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.836136 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.836935 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5d77644dd7-jv6wz" Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.853004 4745 generic.go:334] "Generic (PLEG): container finished" podID="ce71cf3a-de78-452f-b48d-fd8edd70ef35" containerID="2167d93fb5d197090a7afb89b1f6c920a5dcfb7790b293b2d176ca0b2f2e1ad0" exitCode=0 Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.853065 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-644bf58b5b-rhgpb" event={"ID":"ce71cf3a-de78-452f-b48d-fd8edd70ef35","Type":"ContainerDied","Data":"2167d93fb5d197090a7afb89b1f6c920a5dcfb7790b293b2d176ca0b2f2e1ad0"} Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.855510 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-rshbk" event={"ID":"16739af7-770b-4253-be87-ca62b9c21fb9","Type":"ContainerStarted","Data":"62ea94d9b8e4ace99183b3a7039e85d99945afc0610784d58569b0230c933e59"} Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.857362 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-z2v8n" event={"ID":"cc3cdd18-7362-4610-8f9b-7444bd45b09c","Type":"ContainerDied","Data":"3a9373993164362ccbd907a0a0fbf672b8e36f79d061a855e5b5bc7cf434cb2b"} Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.857384 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a9373993164362ccbd907a0a0fbf672b8e36f79d061a855e5b5bc7cf434cb2b" Oct 10 13:34:39 crc kubenswrapper[4745]: I1010 13:34:39.857432 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-z2v8n" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.009562 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dtnk\" (UniqueName: \"kubernetes.io/projected/cc3cdd18-7362-4610-8f9b-7444bd45b09c-kube-api-access-8dtnk\") pod \"cc3cdd18-7362-4610-8f9b-7444bd45b09c\" (UID: \"cc3cdd18-7362-4610-8f9b-7444bd45b09c\") " Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.041193 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc3cdd18-7362-4610-8f9b-7444bd45b09c-kube-api-access-8dtnk" (OuterVolumeSpecName: "kube-api-access-8dtnk") pod "cc3cdd18-7362-4610-8f9b-7444bd45b09c" (UID: "cc3cdd18-7362-4610-8f9b-7444bd45b09c"). InnerVolumeSpecName "kube-api-access-8dtnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.111437 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dtnk\" (UniqueName: \"kubernetes.io/projected/cc3cdd18-7362-4610-8f9b-7444bd45b09c-kube-api-access-8dtnk\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.398469 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.408533 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.519295 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ft74c\" (UniqueName: \"kubernetes.io/projected/ce71cf3a-de78-452f-b48d-fd8edd70ef35-kube-api-access-ft74c\") pod \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.519360 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/604fc859-90dd-4c8b-a216-6009b3dbcbf2-config-data\") pod \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.519449 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-config-data-custom\") pod \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.519480 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-combined-ca-bundle\") pod \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.519505 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-combined-ca-bundle\") pod \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.519550 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce71cf3a-de78-452f-b48d-fd8edd70ef35-logs\") pod \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.519589 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftrmk\" (UniqueName: \"kubernetes.io/projected/604fc859-90dd-4c8b-a216-6009b3dbcbf2-kube-api-access-ftrmk\") pod \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.519610 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/604fc859-90dd-4c8b-a216-6009b3dbcbf2-logs\") pod \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.519635 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/604fc859-90dd-4c8b-a216-6009b3dbcbf2-scripts\") pod \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.519650 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-config-data\") pod \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\" (UID: \"ce71cf3a-de78-452f-b48d-fd8edd70ef35\") " Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.519684 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-horizon-tls-certs\") pod \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.519706 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-horizon-secret-key\") pod \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\" (UID: \"604fc859-90dd-4c8b-a216-6009b3dbcbf2\") " Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.520330 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce71cf3a-de78-452f-b48d-fd8edd70ef35-logs" (OuterVolumeSpecName: "logs") pod "ce71cf3a-de78-452f-b48d-fd8edd70ef35" (UID: "ce71cf3a-de78-452f-b48d-fd8edd70ef35"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.533147 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ce71cf3a-de78-452f-b48d-fd8edd70ef35" (UID: "ce71cf3a-de78-452f-b48d-fd8edd70ef35"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.533578 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/604fc859-90dd-4c8b-a216-6009b3dbcbf2-logs" (OuterVolumeSpecName: "logs") pod "604fc859-90dd-4c8b-a216-6009b3dbcbf2" (UID: "604fc859-90dd-4c8b-a216-6009b3dbcbf2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.535375 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "604fc859-90dd-4c8b-a216-6009b3dbcbf2" (UID: "604fc859-90dd-4c8b-a216-6009b3dbcbf2"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.537434 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce71cf3a-de78-452f-b48d-fd8edd70ef35-kube-api-access-ft74c" (OuterVolumeSpecName: "kube-api-access-ft74c") pod "ce71cf3a-de78-452f-b48d-fd8edd70ef35" (UID: "ce71cf3a-de78-452f-b48d-fd8edd70ef35"). InnerVolumeSpecName "kube-api-access-ft74c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.538545 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/604fc859-90dd-4c8b-a216-6009b3dbcbf2-kube-api-access-ftrmk" (OuterVolumeSpecName: "kube-api-access-ftrmk") pod "604fc859-90dd-4c8b-a216-6009b3dbcbf2" (UID: "604fc859-90dd-4c8b-a216-6009b3dbcbf2"). InnerVolumeSpecName "kube-api-access-ftrmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.580866 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "604fc859-90dd-4c8b-a216-6009b3dbcbf2" (UID: "604fc859-90dd-4c8b-a216-6009b3dbcbf2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.583530 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/604fc859-90dd-4c8b-a216-6009b3dbcbf2-config-data" (OuterVolumeSpecName: "config-data") pod "604fc859-90dd-4c8b-a216-6009b3dbcbf2" (UID: "604fc859-90dd-4c8b-a216-6009b3dbcbf2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.606103 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce71cf3a-de78-452f-b48d-fd8edd70ef35" (UID: "ce71cf3a-de78-452f-b48d-fd8edd70ef35"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.624456 4745 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.624497 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ft74c\" (UniqueName: \"kubernetes.io/projected/ce71cf3a-de78-452f-b48d-fd8edd70ef35-kube-api-access-ft74c\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.624513 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/604fc859-90dd-4c8b-a216-6009b3dbcbf2-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.624525 4745 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.624537 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.624549 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.624563 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce71cf3a-de78-452f-b48d-fd8edd70ef35-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.624575 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftrmk\" (UniqueName: \"kubernetes.io/projected/604fc859-90dd-4c8b-a216-6009b3dbcbf2-kube-api-access-ftrmk\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.624585 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/604fc859-90dd-4c8b-a216-6009b3dbcbf2-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.631991 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "604fc859-90dd-4c8b-a216-6009b3dbcbf2" (UID: "604fc859-90dd-4c8b-a216-6009b3dbcbf2"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.637624 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/604fc859-90dd-4c8b-a216-6009b3dbcbf2-scripts" (OuterVolumeSpecName: "scripts") pod "604fc859-90dd-4c8b-a216-6009b3dbcbf2" (UID: "604fc859-90dd-4c8b-a216-6009b3dbcbf2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.642163 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-config-data" (OuterVolumeSpecName: "config-data") pod "ce71cf3a-de78-452f-b48d-fd8edd70ef35" (UID: "ce71cf3a-de78-452f-b48d-fd8edd70ef35"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.726881 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/604fc859-90dd-4c8b-a216-6009b3dbcbf2-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.726929 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce71cf3a-de78-452f-b48d-fd8edd70ef35-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.726945 4745 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/604fc859-90dd-4c8b-a216-6009b3dbcbf2-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.886199 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c","Type":"ContainerStarted","Data":"53a287959cc291f5d6e381ace206ba788e7ff6e0afcd2a7ee6fc0fdb2a0dfd54"} Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.893317 4745 generic.go:334] "Generic (PLEG): container finished" podID="28904cbe-09fb-499a-8d88-23ac50034c90" containerID="5cee0587bc53b5ae5ba9b9d87902512f80337b576beb7c07a230a232d8f6d114" exitCode=0 Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.893416 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xc9ls" event={"ID":"28904cbe-09fb-499a-8d88-23ac50034c90","Type":"ContainerDied","Data":"5cee0587bc53b5ae5ba9b9d87902512f80337b576beb7c07a230a232d8f6d114"} Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.896301 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-644bf58b5b-rhgpb" event={"ID":"ce71cf3a-de78-452f-b48d-fd8edd70ef35","Type":"ContainerDied","Data":"3063c8dfc65caaa2abcc092ac785436ff0dfec11b71c6f30cbab7fc0e33b9655"} Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.896486 4745 scope.go:117] "RemoveContainer" containerID="2167d93fb5d197090a7afb89b1f6c920a5dcfb7790b293b2d176ca0b2f2e1ad0" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.896677 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-644bf58b5b-rhgpb" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.899639 4745 generic.go:334] "Generic (PLEG): container finished" podID="16739af7-770b-4253-be87-ca62b9c21fb9" containerID="163f31cbd1e3a68a64787aa42c4d8bbff6ba6dc8f43f824d178a92cbea8d0032" exitCode=0 Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.899697 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-rshbk" event={"ID":"16739af7-770b-4253-be87-ca62b9c21fb9","Type":"ContainerDied","Data":"163f31cbd1e3a68a64787aa42c4d8bbff6ba6dc8f43f824d178a92cbea8d0032"} Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.905360 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"5cc91ec2-6a94-4923-82cf-d488c077a97b","Type":"ContainerStarted","Data":"10f3b819915486265c62672eea35cc2ca4f83ef2d1c1304abc776b4d93d235dd"} Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.919767 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65c7d6f5bb-9g4xm" event={"ID":"604fc859-90dd-4c8b-a216-6009b3dbcbf2","Type":"ContainerDied","Data":"32fb177c3021cf6bab2cafe5137093b0a0996835a356486b2465d7ed6093d116"} Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.919891 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65c7d6f5bb-9g4xm" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.926777 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f664e2bc-73b1-464c-acb7-e5ad00a75ec7","Type":"ContainerStarted","Data":"ce5bfcf05221cb458f3d6b9c8c2cf1fec71705e03c71ebe3bbdca92a92bc4c6a"} Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.935988 4745 generic.go:334] "Generic (PLEG): container finished" podID="a3771609-ba86-4413-8cbd-3d266cd0380d" containerID="19b3dcb6e0090f97c5a0c0bbf520c77933b0c453d11c50ece85fea9b057f4eb2" exitCode=0 Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.936459 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-h8ngs" event={"ID":"a3771609-ba86-4413-8cbd-3d266cd0380d","Type":"ContainerDied","Data":"19b3dcb6e0090f97c5a0c0bbf520c77933b0c453d11c50ece85fea9b057f4eb2"} Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.967779 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.847344084 podStartE2EDuration="16.967760081s" podCreationTimestamp="2025-10-10 13:34:24 +0000 UTC" firstStartedPulling="2025-10-10 13:34:25.776207775 +0000 UTC m=+979.673864528" lastFinishedPulling="2025-10-10 13:34:39.896623772 +0000 UTC m=+993.794280525" observedRunningTime="2025-10-10 13:34:40.94170276 +0000 UTC m=+994.839359523" watchObservedRunningTime="2025-10-10 13:34:40.967760081 +0000 UTC m=+994.865416844" Oct 10 13:34:40 crc kubenswrapper[4745]: I1010 13:34:40.998138 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-644bf58b5b-rhgpb"] Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.008998 4745 scope.go:117] "RemoveContainer" containerID="b102f6fab43c0340292ed8434bb7be605f71b341ed968bde26964ddfd2344a59" Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.009020 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-644bf58b5b-rhgpb"] Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.012377 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-65c7d6f5bb-9g4xm"] Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.020556 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-65c7d6f5bb-9g4xm"] Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.039426 4745 scope.go:117] "RemoveContainer" containerID="078e3be7373635f32ef10c0d5768edb4110b552dd87f3f5a0f4b17afe3563529" Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.077993 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-86b9f849df-zktvk" Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.146409 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-748d64ff54-dltxm"] Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.146663 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-748d64ff54-dltxm" podUID="20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb" containerName="neutron-api" containerID="cri-o://9376a83e2ad2b48814abda7d320b0ce3611a5cba3cda60126cab6cee90ecd747" gracePeriod=30 Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.146839 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-748d64ff54-dltxm" podUID="20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb" containerName="neutron-httpd" containerID="cri-o://f7973f4a6d00df3d2f32ea7cc7625ef1ee20cee5b1a1d62752a024768b496050" gracePeriod=30 Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.281134 4745 scope.go:117] "RemoveContainer" containerID="fec5f51d0b50e49ba472342af7f5e04b56f06dcd7a1d7afbd9ec19f6c386fb77" Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.949212 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c","Type":"ContainerStarted","Data":"e96d65833f494c52ef3931066aa73268125155e16629cc5a22dfb2936bb9cac6"} Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.949605 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.949551 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" containerName="cinder-api-log" containerID="cri-o://53a287959cc291f5d6e381ace206ba788e7ff6e0afcd2a7ee6fc0fdb2a0dfd54" gracePeriod=30 Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.949743 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" containerName="cinder-api" containerID="cri-o://e96d65833f494c52ef3931066aa73268125155e16629cc5a22dfb2936bb9cac6" gracePeriod=30 Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.955613 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-rshbk" event={"ID":"16739af7-770b-4253-be87-ca62b9c21fb9","Type":"ContainerStarted","Data":"638df36b11053557be882abd38eed1c627004848d3fe8dd43324de940253460a"} Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.955805 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.958050 4745 generic.go:334] "Generic (PLEG): container finished" podID="20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb" containerID="f7973f4a6d00df3d2f32ea7cc7625ef1ee20cee5b1a1d62752a024768b496050" exitCode=0 Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.958132 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-748d64ff54-dltxm" event={"ID":"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb","Type":"ContainerDied","Data":"f7973f4a6d00df3d2f32ea7cc7625ef1ee20cee5b1a1d62752a024768b496050"} Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.961435 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7560bd3f-0029-4878-bedf-a47020316cca","Type":"ContainerStarted","Data":"63b1748e8b3519cf285aac995f2e25e52ad2b48f8f5e3677bdc9b98524d6d92c"} Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.961494 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7560bd3f-0029-4878-bedf-a47020316cca","Type":"ContainerStarted","Data":"7217da8ac89857a7d10a7d749acbeaff109e593da05c4423ec71081d30e473e8"} Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.977905 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=11.977887166 podStartE2EDuration="11.977887166s" podCreationTimestamp="2025-10-10 13:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:34:41.971961498 +0000 UTC m=+995.869618261" watchObservedRunningTime="2025-10-10 13:34:41.977887166 +0000 UTC m=+995.875543929" Oct 10 13:34:41 crc kubenswrapper[4745]: I1010 13:34:41.995147 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-rshbk" podStartSLOduration=11.995126727 podStartE2EDuration="11.995126727s" podCreationTimestamp="2025-10-10 13:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:34:41.988661656 +0000 UTC m=+995.886318419" watchObservedRunningTime="2025-10-10 13:34:41.995126727 +0000 UTC m=+995.892783490" Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.010447 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=11.091853773 podStartE2EDuration="12.0104297s" podCreationTimestamp="2025-10-10 13:34:30 +0000 UTC" firstStartedPulling="2025-10-10 13:34:39.581559248 +0000 UTC m=+993.479216011" lastFinishedPulling="2025-10-10 13:34:40.500135175 +0000 UTC m=+994.397791938" observedRunningTime="2025-10-10 13:34:42.005207639 +0000 UTC m=+995.902864402" watchObservedRunningTime="2025-10-10 13:34:42.0104297 +0000 UTC m=+995.908086463" Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.333365 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xc9ls" Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.442255 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-h8ngs" Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.463960 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld6qx\" (UniqueName: \"kubernetes.io/projected/28904cbe-09fb-499a-8d88-23ac50034c90-kube-api-access-ld6qx\") pod \"28904cbe-09fb-499a-8d88-23ac50034c90\" (UID: \"28904cbe-09fb-499a-8d88-23ac50034c90\") " Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.470114 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28904cbe-09fb-499a-8d88-23ac50034c90-kube-api-access-ld6qx" (OuterVolumeSpecName: "kube-api-access-ld6qx") pod "28904cbe-09fb-499a-8d88-23ac50034c90" (UID: "28904cbe-09fb-499a-8d88-23ac50034c90"). InnerVolumeSpecName "kube-api-access-ld6qx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.592005 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgzf5\" (UniqueName: \"kubernetes.io/projected/a3771609-ba86-4413-8cbd-3d266cd0380d-kube-api-access-qgzf5\") pod \"a3771609-ba86-4413-8cbd-3d266cd0380d\" (UID: \"a3771609-ba86-4413-8cbd-3d266cd0380d\") " Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.592857 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld6qx\" (UniqueName: \"kubernetes.io/projected/28904cbe-09fb-499a-8d88-23ac50034c90-kube-api-access-ld6qx\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.630123 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3771609-ba86-4413-8cbd-3d266cd0380d-kube-api-access-qgzf5" (OuterVolumeSpecName: "kube-api-access-qgzf5") pod "a3771609-ba86-4413-8cbd-3d266cd0380d" (UID: "a3771609-ba86-4413-8cbd-3d266cd0380d"). InnerVolumeSpecName "kube-api-access-qgzf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.695153 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgzf5\" (UniqueName: \"kubernetes.io/projected/a3771609-ba86-4413-8cbd-3d266cd0380d-kube-api-access-qgzf5\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.764024 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="604fc859-90dd-4c8b-a216-6009b3dbcbf2" path="/var/lib/kubelet/pods/604fc859-90dd-4c8b-a216-6009b3dbcbf2/volumes" Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.764835 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce71cf3a-de78-452f-b48d-fd8edd70ef35" path="/var/lib/kubelet/pods/ce71cf3a-de78-452f-b48d-fd8edd70ef35/volumes" Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.968285 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xc9ls" event={"ID":"28904cbe-09fb-499a-8d88-23ac50034c90","Type":"ContainerDied","Data":"0ae0f2edceb9a91db913ef974d3c44954a48e8aff621a844c5b6c995b581b738"} Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.968320 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ae0f2edceb9a91db913ef974d3c44954a48e8aff621a844c5b6c995b581b738" Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.968386 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xc9ls" Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.970576 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f664e2bc-73b1-464c-acb7-e5ad00a75ec7","Type":"ContainerStarted","Data":"6f397febf2ab4cdfabfe700e0425ad3e72f461a1e97d60b99d9a1793a1d03557"} Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.970683 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerName="ceilometer-central-agent" containerID="cri-o://78e1401b81bb10d24e922e63ca98ca3b70cdb6c3f17f89b5d58cfd6cbf914358" gracePeriod=30 Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.970933 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.971158 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerName="proxy-httpd" containerID="cri-o://6f397febf2ab4cdfabfe700e0425ad3e72f461a1e97d60b99d9a1793a1d03557" gracePeriod=30 Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.971203 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerName="sg-core" containerID="cri-o://ce5bfcf05221cb458f3d6b9c8c2cf1fec71705e03c71ebe3bbdca92a92bc4c6a" gracePeriod=30 Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.971234 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerName="ceilometer-notification-agent" containerID="cri-o://20b58afd107e76758d855767ab2ab220abf53dc8a5540faee28ac0d54f08b0ed" gracePeriod=30 Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.984233 4745 generic.go:334] "Generic (PLEG): container finished" podID="ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" containerID="e96d65833f494c52ef3931066aa73268125155e16629cc5a22dfb2936bb9cac6" exitCode=0 Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.984266 4745 generic.go:334] "Generic (PLEG): container finished" podID="ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" containerID="53a287959cc291f5d6e381ace206ba788e7ff6e0afcd2a7ee6fc0fdb2a0dfd54" exitCode=143 Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.984342 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c","Type":"ContainerDied","Data":"e96d65833f494c52ef3931066aa73268125155e16629cc5a22dfb2936bb9cac6"} Oct 10 13:34:42 crc kubenswrapper[4745]: I1010 13:34:42.984368 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c","Type":"ContainerDied","Data":"53a287959cc291f5d6e381ace206ba788e7ff6e0afcd2a7ee6fc0fdb2a0dfd54"} Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.000006 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-h8ngs" Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.000387 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-h8ngs" event={"ID":"a3771609-ba86-4413-8cbd-3d266cd0380d","Type":"ContainerDied","Data":"4e35f2b2847271b5a77b76833982a0adb24abd370f3311bc1c1483cd7b497716"} Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.000409 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e35f2b2847271b5a77b76833982a0adb24abd370f3311bc1c1483cd7b497716" Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.001783 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.161227963 podStartE2EDuration="15.001764825s" podCreationTimestamp="2025-10-10 13:34:28 +0000 UTC" firstStartedPulling="2025-10-10 13:34:29.361004214 +0000 UTC m=+983.258660997" lastFinishedPulling="2025-10-10 13:34:42.201541096 +0000 UTC m=+996.099197859" observedRunningTime="2025-10-10 13:34:42.993748534 +0000 UTC m=+996.891405307" watchObservedRunningTime="2025-10-10 13:34:43.001764825 +0000 UTC m=+996.899421578" Oct 10 13:34:43 crc kubenswrapper[4745]: E1010 13:34:43.005806 4745 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45009f2e_d0d5_4a20_b62c_d5d0534e1551.slice/crio-a85b3684c58b6faeef3d56c54eb1502b33619172c3103963b89eca08d70b6f4c\": RecentStats: unable to find data in memory cache]" Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.090006 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.201677 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n92qt\" (UniqueName: \"kubernetes.io/projected/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-kube-api-access-n92qt\") pod \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.201826 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-config-data-custom\") pod \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.201857 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-etc-machine-id\") pod \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.201877 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-config-data\") pod \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.201935 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-logs\") pod \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.202015 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-scripts\") pod \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.202037 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-combined-ca-bundle\") pod \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\" (UID: \"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c\") " Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.202810 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" (UID: "ec74267d-39a4-4c0e-a2ff-be5e103e3a9c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.205956 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-logs" (OuterVolumeSpecName: "logs") pod "ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" (UID: "ec74267d-39a4-4c0e-a2ff-be5e103e3a9c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.207875 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-kube-api-access-n92qt" (OuterVolumeSpecName: "kube-api-access-n92qt") pod "ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" (UID: "ec74267d-39a4-4c0e-a2ff-be5e103e3a9c"). InnerVolumeSpecName "kube-api-access-n92qt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.210015 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-scripts" (OuterVolumeSpecName: "scripts") pod "ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" (UID: "ec74267d-39a4-4c0e-a2ff-be5e103e3a9c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.210873 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" (UID: "ec74267d-39a4-4c0e-a2ff-be5e103e3a9c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.231975 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" (UID: "ec74267d-39a4-4c0e-a2ff-be5e103e3a9c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.261395 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-config-data" (OuterVolumeSpecName: "config-data") pod "ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" (UID: "ec74267d-39a4-4c0e-a2ff-be5e103e3a9c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.303771 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.303803 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.303816 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n92qt\" (UniqueName: \"kubernetes.io/projected/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-kube-api-access-n92qt\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.303825 4745 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.303833 4745 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.303842 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:43 crc kubenswrapper[4745]: I1010 13:34:43.303851 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.010912 4745 generic.go:334] "Generic (PLEG): container finished" podID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerID="6f397febf2ab4cdfabfe700e0425ad3e72f461a1e97d60b99d9a1793a1d03557" exitCode=0 Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.011248 4745 generic.go:334] "Generic (PLEG): container finished" podID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerID="ce5bfcf05221cb458f3d6b9c8c2cf1fec71705e03c71ebe3bbdca92a92bc4c6a" exitCode=2 Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.011264 4745 generic.go:334] "Generic (PLEG): container finished" podID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerID="78e1401b81bb10d24e922e63ca98ca3b70cdb6c3f17f89b5d58cfd6cbf914358" exitCode=0 Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.011320 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f664e2bc-73b1-464c-acb7-e5ad00a75ec7","Type":"ContainerDied","Data":"6f397febf2ab4cdfabfe700e0425ad3e72f461a1e97d60b99d9a1793a1d03557"} Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.011354 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f664e2bc-73b1-464c-acb7-e5ad00a75ec7","Type":"ContainerDied","Data":"ce5bfcf05221cb458f3d6b9c8c2cf1fec71705e03c71ebe3bbdca92a92bc4c6a"} Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.011372 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f664e2bc-73b1-464c-acb7-e5ad00a75ec7","Type":"ContainerDied","Data":"78e1401b81bb10d24e922e63ca98ca3b70cdb6c3f17f89b5d58cfd6cbf914358"} Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.015107 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ec74267d-39a4-4c0e-a2ff-be5e103e3a9c","Type":"ContainerDied","Data":"8133ca96d60e4913892109c6d38254b2f4668e24f05bd0d9ef4172bffe93fbbe"} Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.015147 4745 scope.go:117] "RemoveContainer" containerID="e96d65833f494c52ef3931066aa73268125155e16629cc5a22dfb2936bb9cac6" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.015198 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.055153 4745 scope.go:117] "RemoveContainer" containerID="53a287959cc291f5d6e381ace206ba788e7ff6e0afcd2a7ee6fc0fdb2a0dfd54" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.056701 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.065891 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.104902 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 10 13:34:44 crc kubenswrapper[4745]: E1010 13:34:44.105294 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc3cdd18-7362-4610-8f9b-7444bd45b09c" containerName="mariadb-database-create" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105305 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc3cdd18-7362-4610-8f9b-7444bd45b09c" containerName="mariadb-database-create" Oct 10 13:34:44 crc kubenswrapper[4745]: E1010 13:34:44.105315 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" containerName="cinder-api-log" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105321 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" containerName="cinder-api-log" Oct 10 13:34:44 crc kubenswrapper[4745]: E1010 13:34:44.105335 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28904cbe-09fb-499a-8d88-23ac50034c90" containerName="mariadb-database-create" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105342 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="28904cbe-09fb-499a-8d88-23ac50034c90" containerName="mariadb-database-create" Oct 10 13:34:44 crc kubenswrapper[4745]: E1010 13:34:44.105353 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce71cf3a-de78-452f-b48d-fd8edd70ef35" containerName="barbican-api-log" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105359 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce71cf3a-de78-452f-b48d-fd8edd70ef35" containerName="barbican-api-log" Oct 10 13:34:44 crc kubenswrapper[4745]: E1010 13:34:44.105374 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce71cf3a-de78-452f-b48d-fd8edd70ef35" containerName="barbican-api" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105380 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce71cf3a-de78-452f-b48d-fd8edd70ef35" containerName="barbican-api" Oct 10 13:34:44 crc kubenswrapper[4745]: E1010 13:34:44.105394 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="604fc859-90dd-4c8b-a216-6009b3dbcbf2" containerName="horizon-log" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105400 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="604fc859-90dd-4c8b-a216-6009b3dbcbf2" containerName="horizon-log" Oct 10 13:34:44 crc kubenswrapper[4745]: E1010 13:34:44.105413 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" containerName="cinder-api" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105418 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" containerName="cinder-api" Oct 10 13:34:44 crc kubenswrapper[4745]: E1010 13:34:44.105428 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3771609-ba86-4413-8cbd-3d266cd0380d" containerName="mariadb-database-create" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105433 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3771609-ba86-4413-8cbd-3d266cd0380d" containerName="mariadb-database-create" Oct 10 13:34:44 crc kubenswrapper[4745]: E1010 13:34:44.105445 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="604fc859-90dd-4c8b-a216-6009b3dbcbf2" containerName="horizon" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105450 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="604fc859-90dd-4c8b-a216-6009b3dbcbf2" containerName="horizon" Oct 10 13:34:44 crc kubenswrapper[4745]: E1010 13:34:44.105456 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55673aa8-01ac-4a68-b059-cf1a1c0579d6" containerName="init" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105462 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="55673aa8-01ac-4a68-b059-cf1a1c0579d6" containerName="init" Oct 10 13:34:44 crc kubenswrapper[4745]: E1010 13:34:44.105477 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55673aa8-01ac-4a68-b059-cf1a1c0579d6" containerName="dnsmasq-dns" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105483 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="55673aa8-01ac-4a68-b059-cf1a1c0579d6" containerName="dnsmasq-dns" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105643 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" containerName="cinder-api" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105657 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce71cf3a-de78-452f-b48d-fd8edd70ef35" containerName="barbican-api" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105670 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc3cdd18-7362-4610-8f9b-7444bd45b09c" containerName="mariadb-database-create" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105682 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce71cf3a-de78-452f-b48d-fd8edd70ef35" containerName="barbican-api-log" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105692 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="604fc859-90dd-4c8b-a216-6009b3dbcbf2" containerName="horizon" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105706 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" containerName="cinder-api-log" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105713 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3771609-ba86-4413-8cbd-3d266cd0380d" containerName="mariadb-database-create" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105722 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="28904cbe-09fb-499a-8d88-23ac50034c90" containerName="mariadb-database-create" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105743 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="55673aa8-01ac-4a68-b059-cf1a1c0579d6" containerName="dnsmasq-dns" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.105753 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="604fc859-90dd-4c8b-a216-6009b3dbcbf2" containerName="horizon-log" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.106621 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.112223 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.114209 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.114217 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.115560 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.216604 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-config-data\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.216942 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90b61af2-8fcd-40b6-8dc5-92207d219f45-logs\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.217213 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-scripts\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.217357 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/90b61af2-8fcd-40b6-8dc5-92207d219f45-etc-machine-id\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.217482 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-config-data-custom\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.217623 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.217786 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-public-tls-certs\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.217926 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.218050 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5p2l\" (UniqueName: \"kubernetes.io/projected/90b61af2-8fcd-40b6-8dc5-92207d219f45-kube-api-access-x5p2l\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.319310 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90b61af2-8fcd-40b6-8dc5-92207d219f45-logs\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.319421 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-scripts\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.319451 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/90b61af2-8fcd-40b6-8dc5-92207d219f45-etc-machine-id\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.319472 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-config-data-custom\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.319500 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.319532 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-public-tls-certs\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.319546 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.319568 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5p2l\" (UniqueName: \"kubernetes.io/projected/90b61af2-8fcd-40b6-8dc5-92207d219f45-kube-api-access-x5p2l\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.319602 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-config-data\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.320294 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90b61af2-8fcd-40b6-8dc5-92207d219f45-logs\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.321805 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/90b61af2-8fcd-40b6-8dc5-92207d219f45-etc-machine-id\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.325281 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.325315 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.329122 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-scripts\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.329377 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-config-data-custom\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.329562 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-public-tls-certs\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.329799 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90b61af2-8fcd-40b6-8dc5-92207d219f45-config-data\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.340634 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5p2l\" (UniqueName: \"kubernetes.io/projected/90b61af2-8fcd-40b6-8dc5-92207d219f45-kube-api-access-x5p2l\") pod \"cinder-api-0\" (UID: \"90b61af2-8fcd-40b6-8dc5-92207d219f45\") " pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.431222 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.767017 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec74267d-39a4-4c0e-a2ff-be5e103e3a9c" path="/var/lib/kubelet/pods/ec74267d-39a4-4c0e-a2ff-be5e103e3a9c/volumes" Oct 10 13:34:44 crc kubenswrapper[4745]: I1010 13:34:44.906869 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.047664 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f664e2bc-73b1-464c-acb7-e5ad00a75ec7","Type":"ContainerDied","Data":"20b58afd107e76758d855767ab2ab220abf53dc8a5540faee28ac0d54f08b0ed"} Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.051284 4745 generic.go:334] "Generic (PLEG): container finished" podID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerID="20b58afd107e76758d855767ab2ab220abf53dc8a5540faee28ac0d54f08b0ed" exitCode=0 Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.060640 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"90b61af2-8fcd-40b6-8dc5-92207d219f45","Type":"ContainerStarted","Data":"3422a985cd7fc655b394cca76ba0f7a307fd3c5b7d948bbdda8afd7cf97849d6"} Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.405454 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.549183 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-sg-core-conf-yaml\") pod \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.549551 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-log-httpd\") pod \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.549686 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-config-data\") pod \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.549717 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lphmj\" (UniqueName: \"kubernetes.io/projected/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-kube-api-access-lphmj\") pod \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.549794 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-run-httpd\") pod \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.549884 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-combined-ca-bundle\") pod \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.549911 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-scripts\") pod \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\" (UID: \"f664e2bc-73b1-464c-acb7-e5ad00a75ec7\") " Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.550160 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f664e2bc-73b1-464c-acb7-e5ad00a75ec7" (UID: "f664e2bc-73b1-464c-acb7-e5ad00a75ec7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.550255 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f664e2bc-73b1-464c-acb7-e5ad00a75ec7" (UID: "f664e2bc-73b1-464c-acb7-e5ad00a75ec7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.550655 4745 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.550672 4745 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.555819 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-scripts" (OuterVolumeSpecName: "scripts") pod "f664e2bc-73b1-464c-acb7-e5ad00a75ec7" (UID: "f664e2bc-73b1-464c-acb7-e5ad00a75ec7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.564077 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-kube-api-access-lphmj" (OuterVolumeSpecName: "kube-api-access-lphmj") pod "f664e2bc-73b1-464c-acb7-e5ad00a75ec7" (UID: "f664e2bc-73b1-464c-acb7-e5ad00a75ec7"). InnerVolumeSpecName "kube-api-access-lphmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.579856 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f664e2bc-73b1-464c-acb7-e5ad00a75ec7" (UID: "f664e2bc-73b1-464c-acb7-e5ad00a75ec7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.626177 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f664e2bc-73b1-464c-acb7-e5ad00a75ec7" (UID: "f664e2bc-73b1-464c-acb7-e5ad00a75ec7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.654973 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.655008 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.655017 4745 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.655025 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lphmj\" (UniqueName: \"kubernetes.io/projected/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-kube-api-access-lphmj\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.670415 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-config-data" (OuterVolumeSpecName: "config-data") pod "f664e2bc-73b1-464c-acb7-e5ad00a75ec7" (UID: "f664e2bc-73b1-464c-acb7-e5ad00a75ec7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:45 crc kubenswrapper[4745]: I1010 13:34:45.756838 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f664e2bc-73b1-464c-acb7-e5ad00a75ec7-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.103310 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f664e2bc-73b1-464c-acb7-e5ad00a75ec7","Type":"ContainerDied","Data":"b598dbed87005c8bcfe5ba58a857b6243a866d05b2914fdd7ab1b2e0ea2e1948"} Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.103360 4745 scope.go:117] "RemoveContainer" containerID="6f397febf2ab4cdfabfe700e0425ad3e72f461a1e97d60b99d9a1793a1d03557" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.103507 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.148062 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"90b61af2-8fcd-40b6-8dc5-92207d219f45","Type":"ContainerStarted","Data":"9a275b9eec5b6e3173486c1e5872c37d6fecddda988124b76271d0bf65fce7b2"} Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.148395 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.164542 4745 scope.go:117] "RemoveContainer" containerID="ce5bfcf05221cb458f3d6b9c8c2cf1fec71705e03c71ebe3bbdca92a92bc4c6a" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.168465 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.178905 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.182534 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.193013 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:34:46 crc kubenswrapper[4745]: E1010 13:34:46.193379 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerName="proxy-httpd" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.193395 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerName="proxy-httpd" Oct 10 13:34:46 crc kubenswrapper[4745]: E1010 13:34:46.193414 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerName="ceilometer-central-agent" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.193421 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerName="ceilometer-central-agent" Oct 10 13:34:46 crc kubenswrapper[4745]: E1010 13:34:46.193434 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerName="ceilometer-notification-agent" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.193440 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerName="ceilometer-notification-agent" Oct 10 13:34:46 crc kubenswrapper[4745]: E1010 13:34:46.193457 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerName="sg-core" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.193464 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerName="sg-core" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.193616 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerName="proxy-httpd" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.193625 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerName="ceilometer-notification-agent" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.193636 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerName="sg-core" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.193653 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" containerName="ceilometer-central-agent" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.201063 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.225575 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.231538 4745 scope.go:117] "RemoveContainer" containerID="20b58afd107e76758d855767ab2ab220abf53dc8a5540faee28ac0d54f08b0ed" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.232245 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.232472 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.291791 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-9845r"] Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.292151 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" podUID="aedee3a1-1870-4691-a5a4-d5321b90e3a2" containerName="dnsmasq-dns" containerID="cri-o://6fb283d89359918d31200b911555af2620c5f6b69137d2e59804ed5768af383f" gracePeriod=10 Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.297042 4745 scope.go:117] "RemoveContainer" containerID="78e1401b81bb10d24e922e63ca98ca3b70cdb6c3f17f89b5d58cfd6cbf914358" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.372394 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.372623 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-run-httpd\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.372712 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gts5h\" (UniqueName: \"kubernetes.io/projected/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-kube-api-access-gts5h\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.372813 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-scripts\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.373050 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-log-httpd\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.373094 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-config-data\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.373135 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.411067 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.474517 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-log-httpd\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.474559 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-config-data\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.474581 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.474612 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.474655 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-run-httpd\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.474686 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gts5h\" (UniqueName: \"kubernetes.io/projected/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-kube-api-access-gts5h\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.474716 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-scripts\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.475299 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-run-httpd\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.475493 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-log-httpd\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.480658 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.482870 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-scripts\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.484178 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.486411 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-config-data\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.505219 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gts5h\" (UniqueName: \"kubernetes.io/projected/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-kube-api-access-gts5h\") pod \"ceilometer-0\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.569945 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.781109 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f664e2bc-73b1-464c-acb7-e5ad00a75ec7" path="/var/lib/kubelet/pods/f664e2bc-73b1-464c-acb7-e5ad00a75ec7/volumes" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.799002 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.882491 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-dns-svc\") pod \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.882561 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ssz8\" (UniqueName: \"kubernetes.io/projected/aedee3a1-1870-4691-a5a4-d5321b90e3a2-kube-api-access-8ssz8\") pod \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.882618 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-config\") pod \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.882661 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-dns-swift-storage-0\") pod \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.882747 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-ovsdbserver-nb\") pod \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.882773 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-ovsdbserver-sb\") pod \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\" (UID: \"aedee3a1-1870-4691-a5a4-d5321b90e3a2\") " Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.890322 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aedee3a1-1870-4691-a5a4-d5321b90e3a2-kube-api-access-8ssz8" (OuterVolumeSpecName: "kube-api-access-8ssz8") pod "aedee3a1-1870-4691-a5a4-d5321b90e3a2" (UID: "aedee3a1-1870-4691-a5a4-d5321b90e3a2"). InnerVolumeSpecName "kube-api-access-8ssz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.942442 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aedee3a1-1870-4691-a5a4-d5321b90e3a2" (UID: "aedee3a1-1870-4691-a5a4-d5321b90e3a2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.955457 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-config" (OuterVolumeSpecName: "config") pod "aedee3a1-1870-4691-a5a4-d5321b90e3a2" (UID: "aedee3a1-1870-4691-a5a4-d5321b90e3a2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.964173 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aedee3a1-1870-4691-a5a4-d5321b90e3a2" (UID: "aedee3a1-1870-4691-a5a4-d5321b90e3a2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.970133 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aedee3a1-1870-4691-a5a4-d5321b90e3a2" (UID: "aedee3a1-1870-4691-a5a4-d5321b90e3a2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.976020 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "aedee3a1-1870-4691-a5a4-d5321b90e3a2" (UID: "aedee3a1-1870-4691-a5a4-d5321b90e3a2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.984924 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.984970 4745 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.984980 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ssz8\" (UniqueName: \"kubernetes.io/projected/aedee3a1-1870-4691-a5a4-d5321b90e3a2-kube-api-access-8ssz8\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.984989 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.984999 4745 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:46 crc kubenswrapper[4745]: I1010 13:34:46.985007 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aedee3a1-1870-4691-a5a4-d5321b90e3a2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.103006 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.158465 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"90b61af2-8fcd-40b6-8dc5-92207d219f45","Type":"ContainerStarted","Data":"385fbabfbd8e7b9c41ce9319af58fd2132de63ad674cf2d265f7bb3333a6700b"} Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.158612 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.159415 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e","Type":"ContainerStarted","Data":"de8bbaebd370999f0f013c3edf3a505a21c99d41af775d4696711029257a706a"} Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.161211 4745 generic.go:334] "Generic (PLEG): container finished" podID="aedee3a1-1870-4691-a5a4-d5321b90e3a2" containerID="6fb283d89359918d31200b911555af2620c5f6b69137d2e59804ed5768af383f" exitCode=0 Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.161261 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" event={"ID":"aedee3a1-1870-4691-a5a4-d5321b90e3a2","Type":"ContainerDied","Data":"6fb283d89359918d31200b911555af2620c5f6b69137d2e59804ed5768af383f"} Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.161292 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" event={"ID":"aedee3a1-1870-4691-a5a4-d5321b90e3a2","Type":"ContainerDied","Data":"11514f5707a97f5582307623c99788163cc04927c98b858d6ce8c54cda6e99e1"} Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.161312 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-9845r" Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.161324 4745 scope.go:117] "RemoveContainer" containerID="6fb283d89359918d31200b911555af2620c5f6b69137d2e59804ed5768af383f" Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.188463 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.188448785 podStartE2EDuration="3.188448785s" podCreationTimestamp="2025-10-10 13:34:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:34:47.180793703 +0000 UTC m=+1001.078450466" watchObservedRunningTime="2025-10-10 13:34:47.188448785 +0000 UTC m=+1001.086105548" Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.189571 4745 scope.go:117] "RemoveContainer" containerID="2106c08fdc3dac8310716e5f79bc6e7c90283f3b8f4c24b3383765acc57f7386" Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.232364 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.232684 4745 scope.go:117] "RemoveContainer" containerID="6fb283d89359918d31200b911555af2620c5f6b69137d2e59804ed5768af383f" Oct 10 13:34:47 crc kubenswrapper[4745]: E1010 13:34:47.236270 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fb283d89359918d31200b911555af2620c5f6b69137d2e59804ed5768af383f\": container with ID starting with 6fb283d89359918d31200b911555af2620c5f6b69137d2e59804ed5768af383f not found: ID does not exist" containerID="6fb283d89359918d31200b911555af2620c5f6b69137d2e59804ed5768af383f" Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.236316 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fb283d89359918d31200b911555af2620c5f6b69137d2e59804ed5768af383f"} err="failed to get container status \"6fb283d89359918d31200b911555af2620c5f6b69137d2e59804ed5768af383f\": rpc error: code = NotFound desc = could not find container \"6fb283d89359918d31200b911555af2620c5f6b69137d2e59804ed5768af383f\": container with ID starting with 6fb283d89359918d31200b911555af2620c5f6b69137d2e59804ed5768af383f not found: ID does not exist" Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.236342 4745 scope.go:117] "RemoveContainer" containerID="2106c08fdc3dac8310716e5f79bc6e7c90283f3b8f4c24b3383765acc57f7386" Oct 10 13:34:47 crc kubenswrapper[4745]: E1010 13:34:47.240344 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2106c08fdc3dac8310716e5f79bc6e7c90283f3b8f4c24b3383765acc57f7386\": container with ID starting with 2106c08fdc3dac8310716e5f79bc6e7c90283f3b8f4c24b3383765acc57f7386 not found: ID does not exist" containerID="2106c08fdc3dac8310716e5f79bc6e7c90283f3b8f4c24b3383765acc57f7386" Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.240393 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2106c08fdc3dac8310716e5f79bc6e7c90283f3b8f4c24b3383765acc57f7386"} err="failed to get container status \"2106c08fdc3dac8310716e5f79bc6e7c90283f3b8f4c24b3383765acc57f7386\": rpc error: code = NotFound desc = could not find container \"2106c08fdc3dac8310716e5f79bc6e7c90283f3b8f4c24b3383765acc57f7386\": container with ID starting with 2106c08fdc3dac8310716e5f79bc6e7c90283f3b8f4c24b3383765acc57f7386 not found: ID does not exist" Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.252229 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-9845r"] Oct 10 13:34:47 crc kubenswrapper[4745]: I1010 13:34:47.272351 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-9845r"] Oct 10 13:34:48 crc kubenswrapper[4745]: I1010 13:34:48.175157 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e","Type":"ContainerStarted","Data":"77cfacdef3cf2a45a1b677bd0625d791d2fda0d512eff839b0ddfb40ee9d6a1c"} Oct 10 13:34:48 crc kubenswrapper[4745]: I1010 13:34:48.178038 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="7560bd3f-0029-4878-bedf-a47020316cca" containerName="cinder-scheduler" containerID="cri-o://7217da8ac89857a7d10a7d749acbeaff109e593da05c4423ec71081d30e473e8" gracePeriod=30 Oct 10 13:34:48 crc kubenswrapper[4745]: I1010 13:34:48.178450 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="7560bd3f-0029-4878-bedf-a47020316cca" containerName="probe" containerID="cri-o://63b1748e8b3519cf285aac995f2e25e52ad2b48f8f5e3677bdc9b98524d6d92c" gracePeriod=30 Oct 10 13:34:48 crc kubenswrapper[4745]: I1010 13:34:48.479398 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:34:48 crc kubenswrapper[4745]: I1010 13:34:48.479614 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a0048c64-c35e-4358-8e8d-bdb0a65df21c" containerName="glance-log" containerID="cri-o://9daf00e75c695335dcf8fcee3fbe2dc11a393a7a3845077d6f3cf5f5383fe27a" gracePeriod=30 Oct 10 13:34:48 crc kubenswrapper[4745]: I1010 13:34:48.479697 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a0048c64-c35e-4358-8e8d-bdb0a65df21c" containerName="glance-httpd" containerID="cri-o://17e01a5a08b518ada7b6fe6cf0842fb1a9ddeb2c4d93ec40c82f5f5354762385" gracePeriod=30 Oct 10 13:34:48 crc kubenswrapper[4745]: I1010 13:34:48.754767 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aedee3a1-1870-4691-a5a4-d5321b90e3a2" path="/var/lib/kubelet/pods/aedee3a1-1870-4691-a5a4-d5321b90e3a2/volumes" Oct 10 13:34:49 crc kubenswrapper[4745]: I1010 13:34:49.186261 4745 generic.go:334] "Generic (PLEG): container finished" podID="a0048c64-c35e-4358-8e8d-bdb0a65df21c" containerID="9daf00e75c695335dcf8fcee3fbe2dc11a393a7a3845077d6f3cf5f5383fe27a" exitCode=143 Oct 10 13:34:49 crc kubenswrapper[4745]: I1010 13:34:49.186323 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0048c64-c35e-4358-8e8d-bdb0a65df21c","Type":"ContainerDied","Data":"9daf00e75c695335dcf8fcee3fbe2dc11a393a7a3845077d6f3cf5f5383fe27a"} Oct 10 13:34:49 crc kubenswrapper[4745]: I1010 13:34:49.188323 4745 generic.go:334] "Generic (PLEG): container finished" podID="7560bd3f-0029-4878-bedf-a47020316cca" containerID="63b1748e8b3519cf285aac995f2e25e52ad2b48f8f5e3677bdc9b98524d6d92c" exitCode=0 Oct 10 13:34:49 crc kubenswrapper[4745]: I1010 13:34:49.188384 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7560bd3f-0029-4878-bedf-a47020316cca","Type":"ContainerDied","Data":"63b1748e8b3519cf285aac995f2e25e52ad2b48f8f5e3677bdc9b98524d6d92c"} Oct 10 13:34:49 crc kubenswrapper[4745]: I1010 13:34:49.190473 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e","Type":"ContainerStarted","Data":"2b54a0904603f5c65ffd8de935e25f60af0c9c3a91fae496037bcae292a76748"} Oct 10 13:34:49 crc kubenswrapper[4745]: I1010 13:34:49.190496 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e","Type":"ContainerStarted","Data":"b7eeebdff506435ad3bd0575bea7033625902f7f515364ba9d565d65cc22cd70"} Oct 10 13:34:49 crc kubenswrapper[4745]: I1010 13:34:49.617378 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 13:34:49 crc kubenswrapper[4745]: I1010 13:34:49.618032 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ba6456ae-cc8a-45e4-8e01-d37a74e40722" containerName="glance-log" containerID="cri-o://103f7821fe3f99ff187f0620004c15f579d4c341b11b03269580e6565748be2a" gracePeriod=30 Oct 10 13:34:49 crc kubenswrapper[4745]: I1010 13:34:49.618137 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ba6456ae-cc8a-45e4-8e01-d37a74e40722" containerName="glance-httpd" containerID="cri-o://fd34ad7ed94e880a266f45443e59d32a7e8ebfb7c590a7c3bac013fbc000b780" gracePeriod=30 Oct 10 13:34:49 crc kubenswrapper[4745]: I1010 13:34:49.710689 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.210410 4745 generic.go:334] "Generic (PLEG): container finished" podID="ba6456ae-cc8a-45e4-8e01-d37a74e40722" containerID="103f7821fe3f99ff187f0620004c15f579d4c341b11b03269580e6565748be2a" exitCode=143 Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.210744 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ba6456ae-cc8a-45e4-8e01-d37a74e40722","Type":"ContainerDied","Data":"103f7821fe3f99ff187f0620004c15f579d4c341b11b03269580e6565748be2a"} Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.220060 4745 generic.go:334] "Generic (PLEG): container finished" podID="20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb" containerID="9376a83e2ad2b48814abda7d320b0ce3611a5cba3cda60126cab6cee90ecd747" exitCode=0 Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.220242 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-748d64ff54-dltxm" event={"ID":"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb","Type":"ContainerDied","Data":"9376a83e2ad2b48814abda7d320b0ce3611a5cba3cda60126cab6cee90ecd747"} Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.263133 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-3a63-account-create-zprfs"] Oct 10 13:34:50 crc kubenswrapper[4745]: E1010 13:34:50.270291 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aedee3a1-1870-4691-a5a4-d5321b90e3a2" containerName="init" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.270434 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="aedee3a1-1870-4691-a5a4-d5321b90e3a2" containerName="init" Oct 10 13:34:50 crc kubenswrapper[4745]: E1010 13:34:50.270466 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aedee3a1-1870-4691-a5a4-d5321b90e3a2" containerName="dnsmasq-dns" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.270472 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="aedee3a1-1870-4691-a5a4-d5321b90e3a2" containerName="dnsmasq-dns" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.274586 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="aedee3a1-1870-4691-a5a4-d5321b90e3a2" containerName="dnsmasq-dns" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.276200 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3a63-account-create-zprfs" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.282014 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.289480 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmjz8\" (UniqueName: \"kubernetes.io/projected/bff51ec2-99e6-447c-9a70-47da73fc9b5d-kube-api-access-nmjz8\") pod \"nova-api-3a63-account-create-zprfs\" (UID: \"bff51ec2-99e6-447c-9a70-47da73fc9b5d\") " pod="openstack/nova-api-3a63-account-create-zprfs" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.301079 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3a63-account-create-zprfs"] Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.390745 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmjz8\" (UniqueName: \"kubernetes.io/projected/bff51ec2-99e6-447c-9a70-47da73fc9b5d-kube-api-access-nmjz8\") pod \"nova-api-3a63-account-create-zprfs\" (UID: \"bff51ec2-99e6-447c-9a70-47da73fc9b5d\") " pod="openstack/nova-api-3a63-account-create-zprfs" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.412526 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmjz8\" (UniqueName: \"kubernetes.io/projected/bff51ec2-99e6-447c-9a70-47da73fc9b5d-kube-api-access-nmjz8\") pod \"nova-api-3a63-account-create-zprfs\" (UID: \"bff51ec2-99e6-447c-9a70-47da73fc9b5d\") " pod="openstack/nova-api-3a63-account-create-zprfs" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.442490 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-618f-account-create-vhdv7"] Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.443699 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-618f-account-create-vhdv7" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.451948 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.457767 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-618f-account-create-vhdv7"] Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.506327 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.594603 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbcfb\" (UniqueName: \"kubernetes.io/projected/9f7a4122-1442-4714-8556-da06a209febe-kube-api-access-lbcfb\") pod \"nova-cell0-618f-account-create-vhdv7\" (UID: \"9f7a4122-1442-4714-8556-da06a209febe\") " pod="openstack/nova-cell0-618f-account-create-vhdv7" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.612114 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3a63-account-create-zprfs" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.653576 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-3756-account-create-4dd9c"] Oct 10 13:34:50 crc kubenswrapper[4745]: E1010 13:34:50.654199 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb" containerName="neutron-api" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.654266 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb" containerName="neutron-api" Oct 10 13:34:50 crc kubenswrapper[4745]: E1010 13:34:50.654327 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb" containerName="neutron-httpd" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.654387 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb" containerName="neutron-httpd" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.654686 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb" containerName="neutron-api" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.654779 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb" containerName="neutron-httpd" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.655385 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-3756-account-create-4dd9c" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.664150 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.665779 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-3756-account-create-4dd9c"] Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.698953 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-combined-ca-bundle\") pod \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.698992 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-config\") pod \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.699034 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-httpd-config\") pod \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.699050 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkfhd\" (UniqueName: \"kubernetes.io/projected/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-kube-api-access-lkfhd\") pod \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.699103 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-ovndb-tls-certs\") pod \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\" (UID: \"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb\") " Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.699218 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbcfb\" (UniqueName: \"kubernetes.io/projected/9f7a4122-1442-4714-8556-da06a209febe-kube-api-access-lbcfb\") pod \"nova-cell0-618f-account-create-vhdv7\" (UID: \"9f7a4122-1442-4714-8556-da06a209febe\") " pod="openstack/nova-cell0-618f-account-create-vhdv7" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.699260 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9nvv\" (UniqueName: \"kubernetes.io/projected/d41388d5-c639-462f-a5ea-08a27f382511-kube-api-access-l9nvv\") pod \"nova-cell1-3756-account-create-4dd9c\" (UID: \"d41388d5-c639-462f-a5ea-08a27f382511\") " pod="openstack/nova-cell1-3756-account-create-4dd9c" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.704144 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-kube-api-access-lkfhd" (OuterVolumeSpecName: "kube-api-access-lkfhd") pod "20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb" (UID: "20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb"). InnerVolumeSpecName "kube-api-access-lkfhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.715844 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb" (UID: "20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.727481 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbcfb\" (UniqueName: \"kubernetes.io/projected/9f7a4122-1442-4714-8556-da06a209febe-kube-api-access-lbcfb\") pod \"nova-cell0-618f-account-create-vhdv7\" (UID: \"9f7a4122-1442-4714-8556-da06a209febe\") " pod="openstack/nova-cell0-618f-account-create-vhdv7" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.796833 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-config" (OuterVolumeSpecName: "config") pod "20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb" (UID: "20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.804061 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-618f-account-create-vhdv7" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.806370 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9nvv\" (UniqueName: \"kubernetes.io/projected/d41388d5-c639-462f-a5ea-08a27f382511-kube-api-access-l9nvv\") pod \"nova-cell1-3756-account-create-4dd9c\" (UID: \"d41388d5-c639-462f-a5ea-08a27f382511\") " pod="openstack/nova-cell1-3756-account-create-4dd9c" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.806608 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.806674 4745 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.806745 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkfhd\" (UniqueName: \"kubernetes.io/projected/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-kube-api-access-lkfhd\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.818813 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb" (UID: "20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.821378 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9nvv\" (UniqueName: \"kubernetes.io/projected/d41388d5-c639-462f-a5ea-08a27f382511-kube-api-access-l9nvv\") pod \"nova-cell1-3756-account-create-4dd9c\" (UID: \"d41388d5-c639-462f-a5ea-08a27f382511\") " pod="openstack/nova-cell1-3756-account-create-4dd9c" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.854964 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb" (UID: "20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.908660 4745 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:50 crc kubenswrapper[4745]: I1010 13:34:50.908694 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:51 crc kubenswrapper[4745]: I1010 13:34:51.072297 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-3756-account-create-4dd9c" Oct 10 13:34:51 crc kubenswrapper[4745]: I1010 13:34:51.209883 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:51 crc kubenswrapper[4745]: I1010 13:34:51.267656 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-748d64ff54-dltxm" event={"ID":"20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb","Type":"ContainerDied","Data":"8d562c7487b56dce124c35e8e9669ce372d339a4efb7c6150cd4883889b75766"} Oct 10 13:34:51 crc kubenswrapper[4745]: I1010 13:34:51.267714 4745 scope.go:117] "RemoveContainer" containerID="f7973f4a6d00df3d2f32ea7cc7625ef1ee20cee5b1a1d62752a024768b496050" Oct 10 13:34:51 crc kubenswrapper[4745]: I1010 13:34:51.267899 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-748d64ff54-dltxm" Oct 10 13:34:51 crc kubenswrapper[4745]: I1010 13:34:51.304117 4745 scope.go:117] "RemoveContainer" containerID="9376a83e2ad2b48814abda7d320b0ce3611a5cba3cda60126cab6cee90ecd747" Oct 10 13:34:51 crc kubenswrapper[4745]: I1010 13:34:51.322460 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7b8bc58bc6-p44zs" Oct 10 13:34:51 crc kubenswrapper[4745]: I1010 13:34:51.340074 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-748d64ff54-dltxm"] Oct 10 13:34:51 crc kubenswrapper[4745]: I1010 13:34:51.347613 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-748d64ff54-dltxm"] Oct 10 13:34:51 crc kubenswrapper[4745]: I1010 13:34:51.520319 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3a63-account-create-zprfs"] Oct 10 13:34:51 crc kubenswrapper[4745]: I1010 13:34:51.691960 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-618f-account-create-vhdv7"] Oct 10 13:34:51 crc kubenswrapper[4745]: W1010 13:34:51.696135 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f7a4122_1442_4714_8556_da06a209febe.slice/crio-14083ca77fc495a19b6f0fdd652bf5824b7ebaf877899fbe1bcc425217310b75 WatchSource:0}: Error finding container 14083ca77fc495a19b6f0fdd652bf5824b7ebaf877899fbe1bcc425217310b75: Status 404 returned error can't find the container with id 14083ca77fc495a19b6f0fdd652bf5824b7ebaf877899fbe1bcc425217310b75 Oct 10 13:34:51 crc kubenswrapper[4745]: I1010 13:34:51.826629 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-3756-account-create-4dd9c"] Oct 10 13:34:51 crc kubenswrapper[4745]: W1010 13:34:51.867233 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd41388d5_c639_462f_a5ea_08a27f382511.slice/crio-65cad410bdd38f950ddb6ba39218c4970adbc50a30a42483b0b038c8774038ab WatchSource:0}: Error finding container 65cad410bdd38f950ddb6ba39218c4970adbc50a30a42483b0b038c8774038ab: Status 404 returned error can't find the container with id 65cad410bdd38f950ddb6ba39218c4970adbc50a30a42483b0b038c8774038ab Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.277267 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e","Type":"ContainerStarted","Data":"af89481d5714b13b02e6f904cb978eb8398047074b820c3b944fbc05ec6cf395"} Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.277407 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerName="ceilometer-central-agent" containerID="cri-o://77cfacdef3cf2a45a1b677bd0625d791d2fda0d512eff839b0ddfb40ee9d6a1c" gracePeriod=30 Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.277488 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.277471 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerName="proxy-httpd" containerID="cri-o://af89481d5714b13b02e6f904cb978eb8398047074b820c3b944fbc05ec6cf395" gracePeriod=30 Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.277529 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerName="ceilometer-notification-agent" containerID="cri-o://b7eeebdff506435ad3bd0575bea7033625902f7f515364ba9d565d65cc22cd70" gracePeriod=30 Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.277451 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerName="sg-core" containerID="cri-o://2b54a0904603f5c65ffd8de935e25f60af0c9c3a91fae496037bcae292a76748" gracePeriod=30 Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.279089 4745 generic.go:334] "Generic (PLEG): container finished" podID="bff51ec2-99e6-447c-9a70-47da73fc9b5d" containerID="c6590cccecf9b423f770e4917fe5c07b52770199f5d0b44840bfe66534fb9e90" exitCode=0 Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.279142 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3a63-account-create-zprfs" event={"ID":"bff51ec2-99e6-447c-9a70-47da73fc9b5d","Type":"ContainerDied","Data":"c6590cccecf9b423f770e4917fe5c07b52770199f5d0b44840bfe66534fb9e90"} Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.279162 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3a63-account-create-zprfs" event={"ID":"bff51ec2-99e6-447c-9a70-47da73fc9b5d","Type":"ContainerStarted","Data":"989a346728c8d70e251195fa483b082a7e38d5be8d045857ea168b3fd0029af9"} Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.284106 4745 generic.go:334] "Generic (PLEG): container finished" podID="a0048c64-c35e-4358-8e8d-bdb0a65df21c" containerID="17e01a5a08b518ada7b6fe6cf0842fb1a9ddeb2c4d93ec40c82f5f5354762385" exitCode=0 Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.284152 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0048c64-c35e-4358-8e8d-bdb0a65df21c","Type":"ContainerDied","Data":"17e01a5a08b518ada7b6fe6cf0842fb1a9ddeb2c4d93ec40c82f5f5354762385"} Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.284187 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0048c64-c35e-4358-8e8d-bdb0a65df21c","Type":"ContainerDied","Data":"e2a93c49481809c48a060a8bc8aa9f7dea54c56006a4072561abb9ba51960103"} Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.284200 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2a93c49481809c48a060a8bc8aa9f7dea54c56006a4072561abb9ba51960103" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.285650 4745 generic.go:334] "Generic (PLEG): container finished" podID="9f7a4122-1442-4714-8556-da06a209febe" containerID="e674aae9e59433329468c0bf150bafab118433d833fc27a5b1f5fb24a51598a1" exitCode=0 Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.285706 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-618f-account-create-vhdv7" event={"ID":"9f7a4122-1442-4714-8556-da06a209febe","Type":"ContainerDied","Data":"e674aae9e59433329468c0bf150bafab118433d833fc27a5b1f5fb24a51598a1"} Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.285753 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-618f-account-create-vhdv7" event={"ID":"9f7a4122-1442-4714-8556-da06a209febe","Type":"ContainerStarted","Data":"14083ca77fc495a19b6f0fdd652bf5824b7ebaf877899fbe1bcc425217310b75"} Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.288010 4745 generic.go:334] "Generic (PLEG): container finished" podID="d41388d5-c639-462f-a5ea-08a27f382511" containerID="61accf956903ff3eed21347a235a981ff594f9680eb014b7f8c21fa2a0e96599" exitCode=0 Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.288038 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-3756-account-create-4dd9c" event={"ID":"d41388d5-c639-462f-a5ea-08a27f382511","Type":"ContainerDied","Data":"61accf956903ff3eed21347a235a981ff594f9680eb014b7f8c21fa2a0e96599"} Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.288052 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-3756-account-create-4dd9c" event={"ID":"d41388d5-c639-462f-a5ea-08a27f382511","Type":"ContainerStarted","Data":"65cad410bdd38f950ddb6ba39218c4970adbc50a30a42483b0b038c8774038ab"} Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.302792 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.249159004 podStartE2EDuration="6.302771539s" podCreationTimestamp="2025-10-10 13:34:46 +0000 UTC" firstStartedPulling="2025-10-10 13:34:47.102828365 +0000 UTC m=+1001.000485138" lastFinishedPulling="2025-10-10 13:34:51.15644091 +0000 UTC m=+1005.054097673" observedRunningTime="2025-10-10 13:34:52.295385274 +0000 UTC m=+1006.193042047" watchObservedRunningTime="2025-10-10 13:34:52.302771539 +0000 UTC m=+1006.200428302" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.318253 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.333281 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0048c64-c35e-4358-8e8d-bdb0a65df21c-httpd-run\") pod \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.333357 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.333436 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-public-tls-certs\") pod \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.333473 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0048c64-c35e-4358-8e8d-bdb0a65df21c-logs\") pod \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.333498 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k4g9\" (UniqueName: \"kubernetes.io/projected/a0048c64-c35e-4358-8e8d-bdb0a65df21c-kube-api-access-9k4g9\") pod \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.333561 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-combined-ca-bundle\") pod \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.333608 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-scripts\") pod \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.333633 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-config-data\") pod \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\" (UID: \"a0048c64-c35e-4358-8e8d-bdb0a65df21c\") " Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.335490 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0048c64-c35e-4358-8e8d-bdb0a65df21c-logs" (OuterVolumeSpecName: "logs") pod "a0048c64-c35e-4358-8e8d-bdb0a65df21c" (UID: "a0048c64-c35e-4358-8e8d-bdb0a65df21c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.335816 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0048c64-c35e-4358-8e8d-bdb0a65df21c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a0048c64-c35e-4358-8e8d-bdb0a65df21c" (UID: "a0048c64-c35e-4358-8e8d-bdb0a65df21c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.341831 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "a0048c64-c35e-4358-8e8d-bdb0a65df21c" (UID: "a0048c64-c35e-4358-8e8d-bdb0a65df21c"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.346591 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0048c64-c35e-4358-8e8d-bdb0a65df21c-kube-api-access-9k4g9" (OuterVolumeSpecName: "kube-api-access-9k4g9") pod "a0048c64-c35e-4358-8e8d-bdb0a65df21c" (UID: "a0048c64-c35e-4358-8e8d-bdb0a65df21c"). InnerVolumeSpecName "kube-api-access-9k4g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.347878 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-scripts" (OuterVolumeSpecName: "scripts") pod "a0048c64-c35e-4358-8e8d-bdb0a65df21c" (UID: "a0048c64-c35e-4358-8e8d-bdb0a65df21c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.415855 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0048c64-c35e-4358-8e8d-bdb0a65df21c" (UID: "a0048c64-c35e-4358-8e8d-bdb0a65df21c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.424279 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a0048c64-c35e-4358-8e8d-bdb0a65df21c" (UID: "a0048c64-c35e-4358-8e8d-bdb0a65df21c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.424765 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-config-data" (OuterVolumeSpecName: "config-data") pod "a0048c64-c35e-4358-8e8d-bdb0a65df21c" (UID: "a0048c64-c35e-4358-8e8d-bdb0a65df21c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.435221 4745 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.435261 4745 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.435275 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0048c64-c35e-4358-8e8d-bdb0a65df21c-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.435287 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k4g9\" (UniqueName: \"kubernetes.io/projected/a0048c64-c35e-4358-8e8d-bdb0a65df21c-kube-api-access-9k4g9\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.435298 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.435308 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.435318 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0048c64-c35e-4358-8e8d-bdb0a65df21c-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.435328 4745 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0048c64-c35e-4358-8e8d-bdb0a65df21c-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.463435 4745 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.546017 4745 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:52 crc kubenswrapper[4745]: I1010 13:34:52.755452 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb" path="/var/lib/kubelet/pods/20dc0ca6-5d31-4f0e-bf5c-c3fabce95acb/volumes" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.242108 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: E1010 13:34:53.252341 4745 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45009f2e_d0d5_4a20_b62c_d5d0534e1551.slice/crio-a85b3684c58b6faeef3d56c54eb1502b33619172c3103963b89eca08d70b6f4c\": RecentStats: unable to find data in memory cache]" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.256912 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-config-data-custom\") pod \"7560bd3f-0029-4878-bedf-a47020316cca\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.256961 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nl4v\" (UniqueName: \"kubernetes.io/projected/7560bd3f-0029-4878-bedf-a47020316cca-kube-api-access-7nl4v\") pod \"7560bd3f-0029-4878-bedf-a47020316cca\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.257024 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-scripts\") pod \"7560bd3f-0029-4878-bedf-a47020316cca\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.257055 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-combined-ca-bundle\") pod \"7560bd3f-0029-4878-bedf-a47020316cca\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.257110 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-config-data\") pod \"7560bd3f-0029-4878-bedf-a47020316cca\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.257158 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7560bd3f-0029-4878-bedf-a47020316cca-etc-machine-id\") pod \"7560bd3f-0029-4878-bedf-a47020316cca\" (UID: \"7560bd3f-0029-4878-bedf-a47020316cca\") " Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.257488 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7560bd3f-0029-4878-bedf-a47020316cca-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7560bd3f-0029-4878-bedf-a47020316cca" (UID: "7560bd3f-0029-4878-bedf-a47020316cca"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.270827 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7560bd3f-0029-4878-bedf-a47020316cca-kube-api-access-7nl4v" (OuterVolumeSpecName: "kube-api-access-7nl4v") pod "7560bd3f-0029-4878-bedf-a47020316cca" (UID: "7560bd3f-0029-4878-bedf-a47020316cca"). InnerVolumeSpecName "kube-api-access-7nl4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.284134 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7560bd3f-0029-4878-bedf-a47020316cca" (UID: "7560bd3f-0029-4878-bedf-a47020316cca"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.292574 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-scripts" (OuterVolumeSpecName: "scripts") pod "7560bd3f-0029-4878-bedf-a47020316cca" (UID: "7560bd3f-0029-4878-bedf-a47020316cca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.310831 4745 generic.go:334] "Generic (PLEG): container finished" podID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerID="af89481d5714b13b02e6f904cb978eb8398047074b820c3b944fbc05ec6cf395" exitCode=0 Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.310860 4745 generic.go:334] "Generic (PLEG): container finished" podID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerID="2b54a0904603f5c65ffd8de935e25f60af0c9c3a91fae496037bcae292a76748" exitCode=2 Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.310870 4745 generic.go:334] "Generic (PLEG): container finished" podID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerID="b7eeebdff506435ad3bd0575bea7033625902f7f515364ba9d565d65cc22cd70" exitCode=0 Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.310923 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e","Type":"ContainerDied","Data":"af89481d5714b13b02e6f904cb978eb8398047074b820c3b944fbc05ec6cf395"} Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.310948 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e","Type":"ContainerDied","Data":"2b54a0904603f5c65ffd8de935e25f60af0c9c3a91fae496037bcae292a76748"} Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.310959 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e","Type":"ContainerDied","Data":"b7eeebdff506435ad3bd0575bea7033625902f7f515364ba9d565d65cc22cd70"} Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.316813 4745 generic.go:334] "Generic (PLEG): container finished" podID="ba6456ae-cc8a-45e4-8e01-d37a74e40722" containerID="fd34ad7ed94e880a266f45443e59d32a7e8ebfb7c590a7c3bac013fbc000b780" exitCode=0 Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.316878 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ba6456ae-cc8a-45e4-8e01-d37a74e40722","Type":"ContainerDied","Data":"fd34ad7ed94e880a266f45443e59d32a7e8ebfb7c590a7c3bac013fbc000b780"} Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.319017 4745 generic.go:334] "Generic (PLEG): container finished" podID="7560bd3f-0029-4878-bedf-a47020316cca" containerID="7217da8ac89857a7d10a7d749acbeaff109e593da05c4423ec71081d30e473e8" exitCode=0 Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.319090 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.319084 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7560bd3f-0029-4878-bedf-a47020316cca","Type":"ContainerDied","Data":"7217da8ac89857a7d10a7d749acbeaff109e593da05c4423ec71081d30e473e8"} Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.319436 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7560bd3f-0029-4878-bedf-a47020316cca","Type":"ContainerDied","Data":"3d6f561ffeb18c437cfef483d4d72dd510b62a3cbecfe016b0beaf750dc043e2"} Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.319454 4745 scope.go:117] "RemoveContainer" containerID="63b1748e8b3519cf285aac995f2e25e52ad2b48f8f5e3677bdc9b98524d6d92c" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.320790 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.358705 4745 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7560bd3f-0029-4878-bedf-a47020316cca-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.358749 4745 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.358759 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nl4v\" (UniqueName: \"kubernetes.io/projected/7560bd3f-0029-4878-bedf-a47020316cca-kube-api-access-7nl4v\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.358768 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.376171 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.381287 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.382128 4745 scope.go:117] "RemoveContainer" containerID="7217da8ac89857a7d10a7d749acbeaff109e593da05c4423ec71081d30e473e8" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.383435 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7560bd3f-0029-4878-bedf-a47020316cca" (UID: "7560bd3f-0029-4878-bedf-a47020316cca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.389445 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:34:53 crc kubenswrapper[4745]: E1010 13:34:53.389839 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0048c64-c35e-4358-8e8d-bdb0a65df21c" containerName="glance-httpd" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.389854 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0048c64-c35e-4358-8e8d-bdb0a65df21c" containerName="glance-httpd" Oct 10 13:34:53 crc kubenswrapper[4745]: E1010 13:34:53.389868 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7560bd3f-0029-4878-bedf-a47020316cca" containerName="cinder-scheduler" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.389875 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="7560bd3f-0029-4878-bedf-a47020316cca" containerName="cinder-scheduler" Oct 10 13:34:53 crc kubenswrapper[4745]: E1010 13:34:53.389905 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0048c64-c35e-4358-8e8d-bdb0a65df21c" containerName="glance-log" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.389911 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0048c64-c35e-4358-8e8d-bdb0a65df21c" containerName="glance-log" Oct 10 13:34:53 crc kubenswrapper[4745]: E1010 13:34:53.389926 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7560bd3f-0029-4878-bedf-a47020316cca" containerName="probe" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.389932 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="7560bd3f-0029-4878-bedf-a47020316cca" containerName="probe" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.390086 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="7560bd3f-0029-4878-bedf-a47020316cca" containerName="probe" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.390096 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0048c64-c35e-4358-8e8d-bdb0a65df21c" containerName="glance-httpd" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.390110 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0048c64-c35e-4358-8e8d-bdb0a65df21c" containerName="glance-log" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.390117 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="7560bd3f-0029-4878-bedf-a47020316cca" containerName="cinder-scheduler" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.391530 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.398262 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.398451 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.412929 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.454043 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.457298 4745 scope.go:117] "RemoveContainer" containerID="63b1748e8b3519cf285aac995f2e25e52ad2b48f8f5e3677bdc9b98524d6d92c" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.459619 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btp2g\" (UniqueName: \"kubernetes.io/projected/ba6456ae-cc8a-45e4-8e01-d37a74e40722-kube-api-access-btp2g\") pod \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.459665 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.459705 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-combined-ca-bundle\") pod \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.459757 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-internal-tls-certs\") pod \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.459787 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ba6456ae-cc8a-45e4-8e01-d37a74e40722-httpd-run\") pod \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.459816 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-scripts\") pod \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.459830 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-config-data\") pod \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.459851 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba6456ae-cc8a-45e4-8e01-d37a74e40722-logs\") pod \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\" (UID: \"ba6456ae-cc8a-45e4-8e01-d37a74e40722\") " Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.459986 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-logs\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.460009 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.460061 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-scripts\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.460076 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.460099 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-config-data\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.460113 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzb4v\" (UniqueName: \"kubernetes.io/projected/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-kube-api-access-tzb4v\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.460159 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.460184 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.460229 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.461368 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba6456ae-cc8a-45e4-8e01-d37a74e40722-logs" (OuterVolumeSpecName: "logs") pod "ba6456ae-cc8a-45e4-8e01-d37a74e40722" (UID: "ba6456ae-cc8a-45e4-8e01-d37a74e40722"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.464195 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba6456ae-cc8a-45e4-8e01-d37a74e40722-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ba6456ae-cc8a-45e4-8e01-d37a74e40722" (UID: "ba6456ae-cc8a-45e4-8e01-d37a74e40722"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:34:53 crc kubenswrapper[4745]: E1010 13:34:53.464329 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63b1748e8b3519cf285aac995f2e25e52ad2b48f8f5e3677bdc9b98524d6d92c\": container with ID starting with 63b1748e8b3519cf285aac995f2e25e52ad2b48f8f5e3677bdc9b98524d6d92c not found: ID does not exist" containerID="63b1748e8b3519cf285aac995f2e25e52ad2b48f8f5e3677bdc9b98524d6d92c" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.464363 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63b1748e8b3519cf285aac995f2e25e52ad2b48f8f5e3677bdc9b98524d6d92c"} err="failed to get container status \"63b1748e8b3519cf285aac995f2e25e52ad2b48f8f5e3677bdc9b98524d6d92c\": rpc error: code = NotFound desc = could not find container \"63b1748e8b3519cf285aac995f2e25e52ad2b48f8f5e3677bdc9b98524d6d92c\": container with ID starting with 63b1748e8b3519cf285aac995f2e25e52ad2b48f8f5e3677bdc9b98524d6d92c not found: ID does not exist" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.464389 4745 scope.go:117] "RemoveContainer" containerID="7217da8ac89857a7d10a7d749acbeaff109e593da05c4423ec71081d30e473e8" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.466560 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba6456ae-cc8a-45e4-8e01-d37a74e40722-kube-api-access-btp2g" (OuterVolumeSpecName: "kube-api-access-btp2g") pod "ba6456ae-cc8a-45e4-8e01-d37a74e40722" (UID: "ba6456ae-cc8a-45e4-8e01-d37a74e40722"). InnerVolumeSpecName "kube-api-access-btp2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.473028 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-scripts" (OuterVolumeSpecName: "scripts") pod "ba6456ae-cc8a-45e4-8e01-d37a74e40722" (UID: "ba6456ae-cc8a-45e4-8e01-d37a74e40722"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:53 crc kubenswrapper[4745]: E1010 13:34:53.492003 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7217da8ac89857a7d10a7d749acbeaff109e593da05c4423ec71081d30e473e8\": container with ID starting with 7217da8ac89857a7d10a7d749acbeaff109e593da05c4423ec71081d30e473e8 not found: ID does not exist" containerID="7217da8ac89857a7d10a7d749acbeaff109e593da05c4423ec71081d30e473e8" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.492044 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7217da8ac89857a7d10a7d749acbeaff109e593da05c4423ec71081d30e473e8"} err="failed to get container status \"7217da8ac89857a7d10a7d749acbeaff109e593da05c4423ec71081d30e473e8\": rpc error: code = NotFound desc = could not find container \"7217da8ac89857a7d10a7d749acbeaff109e593da05c4423ec71081d30e473e8\": container with ID starting with 7217da8ac89857a7d10a7d749acbeaff109e593da05c4423ec71081d30e473e8 not found: ID does not exist" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.492100 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-config-data" (OuterVolumeSpecName: "config-data") pod "7560bd3f-0029-4878-bedf-a47020316cca" (UID: "7560bd3f-0029-4878-bedf-a47020316cca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.507268 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "ba6456ae-cc8a-45e4-8e01-d37a74e40722" (UID: "ba6456ae-cc8a-45e4-8e01-d37a74e40722"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.536001 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ba6456ae-cc8a-45e4-8e01-d37a74e40722" (UID: "ba6456ae-cc8a-45e4-8e01-d37a74e40722"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.555632 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ba6456ae-cc8a-45e4-8e01-d37a74e40722" (UID: "ba6456ae-cc8a-45e4-8e01-d37a74e40722"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.557548 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-config-data" (OuterVolumeSpecName: "config-data") pod "ba6456ae-cc8a-45e4-8e01-d37a74e40722" (UID: "ba6456ae-cc8a-45e4-8e01-d37a74e40722"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.560964 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-scripts\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.560999 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.561029 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzb4v\" (UniqueName: \"kubernetes.io/projected/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-kube-api-access-tzb4v\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.561044 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-config-data\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.561092 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.561119 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.561152 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-logs\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.561169 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.561236 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7560bd3f-0029-4878-bedf-a47020316cca-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.561249 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btp2g\" (UniqueName: \"kubernetes.io/projected/ba6456ae-cc8a-45e4-8e01-d37a74e40722-kube-api-access-btp2g\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.561268 4745 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.561277 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.561287 4745 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.561297 4745 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ba6456ae-cc8a-45e4-8e01-d37a74e40722-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.561305 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.561313 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba6456ae-cc8a-45e4-8e01-d37a74e40722-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.561320 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba6456ae-cc8a-45e4-8e01-d37a74e40722-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.567935 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-logs\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.568018 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.568696 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.568984 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-scripts\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.571776 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.579567 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.579668 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-config-data\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.596492 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzb4v\" (UniqueName: \"kubernetes.io/projected/d2ba148a-0fdd-45d3-91e7-106c98cc7d2e-kube-api-access-tzb4v\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.600039 4745 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.668351 4745 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.682640 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e\") " pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.711176 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.722131 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.727261 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.728894 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 13:34:53 crc kubenswrapper[4745]: E1010 13:34:53.729456 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba6456ae-cc8a-45e4-8e01-d37a74e40722" containerName="glance-log" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.729523 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba6456ae-cc8a-45e4-8e01-d37a74e40722" containerName="glance-log" Oct 10 13:34:53 crc kubenswrapper[4745]: E1010 13:34:53.729571 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba6456ae-cc8a-45e4-8e01-d37a74e40722" containerName="glance-httpd" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.729578 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba6456ae-cc8a-45e4-8e01-d37a74e40722" containerName="glance-httpd" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.729828 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba6456ae-cc8a-45e4-8e01-d37a74e40722" containerName="glance-log" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.729860 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba6456ae-cc8a-45e4-8e01-d37a74e40722" containerName="glance-httpd" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.747268 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.747423 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.748991 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.752780 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3a63-account-create-zprfs" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.878652 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmjz8\" (UniqueName: \"kubernetes.io/projected/bff51ec2-99e6-447c-9a70-47da73fc9b5d-kube-api-access-nmjz8\") pod \"bff51ec2-99e6-447c-9a70-47da73fc9b5d\" (UID: \"bff51ec2-99e6-447c-9a70-47da73fc9b5d\") " Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.879333 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513c43ae-c96a-4ef0-8162-708490db52d6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.879373 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc8k6\" (UniqueName: \"kubernetes.io/projected/513c43ae-c96a-4ef0-8162-708490db52d6-kube-api-access-hc8k6\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.879399 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/513c43ae-c96a-4ef0-8162-708490db52d6-scripts\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.879427 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/513c43ae-c96a-4ef0-8162-708490db52d6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.879447 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/513c43ae-c96a-4ef0-8162-708490db52d6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.879524 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/513c43ae-c96a-4ef0-8162-708490db52d6-config-data\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.895358 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bff51ec2-99e6-447c-9a70-47da73fc9b5d-kube-api-access-nmjz8" (OuterVolumeSpecName: "kube-api-access-nmjz8") pod "bff51ec2-99e6-447c-9a70-47da73fc9b5d" (UID: "bff51ec2-99e6-447c-9a70-47da73fc9b5d"). InnerVolumeSpecName "kube-api-access-nmjz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.927530 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-3756-account-create-4dd9c" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.928768 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-618f-account-create-vhdv7" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.981038 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc8k6\" (UniqueName: \"kubernetes.io/projected/513c43ae-c96a-4ef0-8162-708490db52d6-kube-api-access-hc8k6\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.981132 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/513c43ae-c96a-4ef0-8162-708490db52d6-scripts\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.981185 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/513c43ae-c96a-4ef0-8162-708490db52d6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.981225 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/513c43ae-c96a-4ef0-8162-708490db52d6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.981339 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/513c43ae-c96a-4ef0-8162-708490db52d6-config-data\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.981499 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513c43ae-c96a-4ef0-8162-708490db52d6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.981618 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmjz8\" (UniqueName: \"kubernetes.io/projected/bff51ec2-99e6-447c-9a70-47da73fc9b5d-kube-api-access-nmjz8\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.981630 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/513c43ae-c96a-4ef0-8162-708490db52d6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.987381 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/513c43ae-c96a-4ef0-8162-708490db52d6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.987583 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/513c43ae-c96a-4ef0-8162-708490db52d6-config-data\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.991255 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/513c43ae-c96a-4ef0-8162-708490db52d6-scripts\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.994408 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513c43ae-c96a-4ef0-8162-708490db52d6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:53 crc kubenswrapper[4745]: I1010 13:34:53.996546 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc8k6\" (UniqueName: \"kubernetes.io/projected/513c43ae-c96a-4ef0-8162-708490db52d6-kube-api-access-hc8k6\") pod \"cinder-scheduler-0\" (UID: \"513c43ae-c96a-4ef0-8162-708490db52d6\") " pod="openstack/cinder-scheduler-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.083462 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9nvv\" (UniqueName: \"kubernetes.io/projected/d41388d5-c639-462f-a5ea-08a27f382511-kube-api-access-l9nvv\") pod \"d41388d5-c639-462f-a5ea-08a27f382511\" (UID: \"d41388d5-c639-462f-a5ea-08a27f382511\") " Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.083555 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbcfb\" (UniqueName: \"kubernetes.io/projected/9f7a4122-1442-4714-8556-da06a209febe-kube-api-access-lbcfb\") pod \"9f7a4122-1442-4714-8556-da06a209febe\" (UID: \"9f7a4122-1442-4714-8556-da06a209febe\") " Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.086753 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d41388d5-c639-462f-a5ea-08a27f382511-kube-api-access-l9nvv" (OuterVolumeSpecName: "kube-api-access-l9nvv") pod "d41388d5-c639-462f-a5ea-08a27f382511" (UID: "d41388d5-c639-462f-a5ea-08a27f382511"). InnerVolumeSpecName "kube-api-access-l9nvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.088633 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f7a4122-1442-4714-8556-da06a209febe-kube-api-access-lbcfb" (OuterVolumeSpecName: "kube-api-access-lbcfb") pod "9f7a4122-1442-4714-8556-da06a209febe" (UID: "9f7a4122-1442-4714-8556-da06a209febe"). InnerVolumeSpecName "kube-api-access-lbcfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.095038 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.185848 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbcfb\" (UniqueName: \"kubernetes.io/projected/9f7a4122-1442-4714-8556-da06a209febe-kube-api-access-lbcfb\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.185874 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9nvv\" (UniqueName: \"kubernetes.io/projected/d41388d5-c639-462f-a5ea-08a27f382511-kube-api-access-l9nvv\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.333188 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ba6456ae-cc8a-45e4-8e01-d37a74e40722","Type":"ContainerDied","Data":"f43ce3d644d4b3802aaf005bebff396f83f23050492aaa9f4a5c76c2870c110b"} Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.333482 4745 scope.go:117] "RemoveContainer" containerID="fd34ad7ed94e880a266f45443e59d32a7e8ebfb7c590a7c3bac013fbc000b780" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.333251 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.337287 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3a63-account-create-zprfs" event={"ID":"bff51ec2-99e6-447c-9a70-47da73fc9b5d","Type":"ContainerDied","Data":"989a346728c8d70e251195fa483b082a7e38d5be8d045857ea168b3fd0029af9"} Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.337314 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="989a346728c8d70e251195fa483b082a7e38d5be8d045857ea168b3fd0029af9" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.337343 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3a63-account-create-zprfs" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.343791 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-618f-account-create-vhdv7" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.352920 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-618f-account-create-vhdv7" event={"ID":"9f7a4122-1442-4714-8556-da06a209febe","Type":"ContainerDied","Data":"14083ca77fc495a19b6f0fdd652bf5824b7ebaf877899fbe1bcc425217310b75"} Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.352963 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14083ca77fc495a19b6f0fdd652bf5824b7ebaf877899fbe1bcc425217310b75" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.357632 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-3756-account-create-4dd9c" event={"ID":"d41388d5-c639-462f-a5ea-08a27f382511","Type":"ContainerDied","Data":"65cad410bdd38f950ddb6ba39218c4970adbc50a30a42483b0b038c8774038ab"} Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.357671 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65cad410bdd38f950ddb6ba39218c4970adbc50a30a42483b0b038c8774038ab" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.357722 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-3756-account-create-4dd9c" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.360298 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.375379 4745 scope.go:117] "RemoveContainer" containerID="103f7821fe3f99ff187f0620004c15f579d4c341b11b03269580e6565748be2a" Oct 10 13:34:54 crc kubenswrapper[4745]: W1010 13:34:54.381169 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2ba148a_0fdd_45d3_91e7_106c98cc7d2e.slice/crio-f2efb19f6f89a1b4c68d37292e45c37d3d6e4c7327d5b1e212bca91ec52d06be WatchSource:0}: Error finding container f2efb19f6f89a1b4c68d37292e45c37d3d6e4c7327d5b1e212bca91ec52d06be: Status 404 returned error can't find the container with id f2efb19f6f89a1b4c68d37292e45c37d3d6e4c7327d5b1e212bca91ec52d06be Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.403777 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.414637 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.436292 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 13:34:54 crc kubenswrapper[4745]: E1010 13:34:54.436665 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d41388d5-c639-462f-a5ea-08a27f382511" containerName="mariadb-account-create" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.436681 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="d41388d5-c639-462f-a5ea-08a27f382511" containerName="mariadb-account-create" Oct 10 13:34:54 crc kubenswrapper[4745]: E1010 13:34:54.436702 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bff51ec2-99e6-447c-9a70-47da73fc9b5d" containerName="mariadb-account-create" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.436709 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="bff51ec2-99e6-447c-9a70-47da73fc9b5d" containerName="mariadb-account-create" Oct 10 13:34:54 crc kubenswrapper[4745]: E1010 13:34:54.436720 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f7a4122-1442-4714-8556-da06a209febe" containerName="mariadb-account-create" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.436739 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f7a4122-1442-4714-8556-da06a209febe" containerName="mariadb-account-create" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.437668 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="bff51ec2-99e6-447c-9a70-47da73fc9b5d" containerName="mariadb-account-create" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.437694 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f7a4122-1442-4714-8556-da06a209febe" containerName="mariadb-account-create" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.437710 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="d41388d5-c639-462f-a5ea-08a27f382511" containerName="mariadb-account-create" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.438585 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.440819 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.443254 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.467920 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.523206 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 13:34:54 crc kubenswrapper[4745]: W1010 13:34:54.541089 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod513c43ae_c96a_4ef0_8162_708490db52d6.slice/crio-f3efc92b74be9d1fd1a792454cb49dd6f4430e11d9a0f3c065eb06e04f797e58 WatchSource:0}: Error finding container f3efc92b74be9d1fd1a792454cb49dd6f4430e11d9a0f3c065eb06e04f797e58: Status 404 returned error can't find the container with id f3efc92b74be9d1fd1a792454cb49dd6f4430e11d9a0f3c065eb06e04f797e58 Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.592012 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.592261 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a7607af-7365-45dc-9a8e-91719c20cdcf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.592297 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a7607af-7365-45dc-9a8e-91719c20cdcf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.592334 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0a7607af-7365-45dc-9a8e-91719c20cdcf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.592374 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a7607af-7365-45dc-9a8e-91719c20cdcf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.592413 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a7607af-7365-45dc-9a8e-91719c20cdcf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.592454 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a7607af-7365-45dc-9a8e-91719c20cdcf-logs\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.592598 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4h69\" (UniqueName: \"kubernetes.io/projected/0a7607af-7365-45dc-9a8e-91719c20cdcf-kube-api-access-s4h69\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.693900 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a7607af-7365-45dc-9a8e-91719c20cdcf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.693974 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0a7607af-7365-45dc-9a8e-91719c20cdcf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.694006 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a7607af-7365-45dc-9a8e-91719c20cdcf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.694048 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a7607af-7365-45dc-9a8e-91719c20cdcf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.694088 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a7607af-7365-45dc-9a8e-91719c20cdcf-logs\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.694134 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4h69\" (UniqueName: \"kubernetes.io/projected/0a7607af-7365-45dc-9a8e-91719c20cdcf-kube-api-access-s4h69\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.694188 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.694221 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a7607af-7365-45dc-9a8e-91719c20cdcf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.695664 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.695688 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a7607af-7365-45dc-9a8e-91719c20cdcf-logs\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.695954 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0a7607af-7365-45dc-9a8e-91719c20cdcf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.699271 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a7607af-7365-45dc-9a8e-91719c20cdcf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.699795 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a7607af-7365-45dc-9a8e-91719c20cdcf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.700199 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a7607af-7365-45dc-9a8e-91719c20cdcf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.703296 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a7607af-7365-45dc-9a8e-91719c20cdcf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.720130 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4h69\" (UniqueName: \"kubernetes.io/projected/0a7607af-7365-45dc-9a8e-91719c20cdcf-kube-api-access-s4h69\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.742463 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"0a7607af-7365-45dc-9a8e-91719c20cdcf\") " pod="openstack/glance-default-internal-api-0" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.759836 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7560bd3f-0029-4878-bedf-a47020316cca" path="/var/lib/kubelet/pods/7560bd3f-0029-4878-bedf-a47020316cca/volumes" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.764130 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0048c64-c35e-4358-8e8d-bdb0a65df21c" path="/var/lib/kubelet/pods/a0048c64-c35e-4358-8e8d-bdb0a65df21c/volumes" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.764841 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba6456ae-cc8a-45e4-8e01-d37a74e40722" path="/var/lib/kubelet/pods/ba6456ae-cc8a-45e4-8e01-d37a74e40722/volumes" Oct 10 13:34:54 crc kubenswrapper[4745]: I1010 13:34:54.776663 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.372564 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.379390 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e","Type":"ContainerStarted","Data":"452e81535771969c081e419e473041c2f3adf6c52daf6bf34a7165fdeaa4d4b5"} Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.379433 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e","Type":"ContainerStarted","Data":"f2efb19f6f89a1b4c68d37292e45c37d3d6e4c7327d5b1e212bca91ec52d06be"} Oct 10 13:34:55 crc kubenswrapper[4745]: W1010 13:34:55.380184 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a7607af_7365_45dc_9a8e_91719c20cdcf.slice/crio-ecd7d756e9c16aa0725e231c6eb76a740119ae14fcf145c620f089212f91500e WatchSource:0}: Error finding container ecd7d756e9c16aa0725e231c6eb76a740119ae14fcf145c620f089212f91500e: Status 404 returned error can't find the container with id ecd7d756e9c16aa0725e231c6eb76a740119ae14fcf145c620f089212f91500e Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.380779 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"513c43ae-c96a-4ef0-8162-708490db52d6","Type":"ContainerStarted","Data":"47d84b52304aadf98f4cf36d607e1b2f0d8e9218b4ad16c9a715d320b93e5bcd"} Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.380803 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"513c43ae-c96a-4ef0-8162-708490db52d6","Type":"ContainerStarted","Data":"f3efc92b74be9d1fd1a792454cb49dd6f4430e11d9a0f3c065eb06e04f797e58"} Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.690438 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wp869"] Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.692566 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wp869" Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.700029 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.700411 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.700604 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bx7wb" Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.712273 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wp869"] Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.819386 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-config-data\") pod \"nova-cell0-conductor-db-sync-wp869\" (UID: \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\") " pod="openstack/nova-cell0-conductor-db-sync-wp869" Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.819443 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjh9m\" (UniqueName: \"kubernetes.io/projected/907a33f4-60a8-44b7-8b1d-5fb979a84e62-kube-api-access-vjh9m\") pod \"nova-cell0-conductor-db-sync-wp869\" (UID: \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\") " pod="openstack/nova-cell0-conductor-db-sync-wp869" Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.819480 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wp869\" (UID: \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\") " pod="openstack/nova-cell0-conductor-db-sync-wp869" Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.819510 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-scripts\") pod \"nova-cell0-conductor-db-sync-wp869\" (UID: \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\") " pod="openstack/nova-cell0-conductor-db-sync-wp869" Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.923315 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-config-data\") pod \"nova-cell0-conductor-db-sync-wp869\" (UID: \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\") " pod="openstack/nova-cell0-conductor-db-sync-wp869" Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.923364 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjh9m\" (UniqueName: \"kubernetes.io/projected/907a33f4-60a8-44b7-8b1d-5fb979a84e62-kube-api-access-vjh9m\") pod \"nova-cell0-conductor-db-sync-wp869\" (UID: \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\") " pod="openstack/nova-cell0-conductor-db-sync-wp869" Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.923405 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wp869\" (UID: \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\") " pod="openstack/nova-cell0-conductor-db-sync-wp869" Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.923442 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-scripts\") pod \"nova-cell0-conductor-db-sync-wp869\" (UID: \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\") " pod="openstack/nova-cell0-conductor-db-sync-wp869" Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.928420 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-scripts\") pod \"nova-cell0-conductor-db-sync-wp869\" (UID: \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\") " pod="openstack/nova-cell0-conductor-db-sync-wp869" Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.930517 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-config-data\") pod \"nova-cell0-conductor-db-sync-wp869\" (UID: \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\") " pod="openstack/nova-cell0-conductor-db-sync-wp869" Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.934277 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wp869\" (UID: \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\") " pod="openstack/nova-cell0-conductor-db-sync-wp869" Oct 10 13:34:55 crc kubenswrapper[4745]: I1010 13:34:55.938138 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjh9m\" (UniqueName: \"kubernetes.io/projected/907a33f4-60a8-44b7-8b1d-5fb979a84e62-kube-api-access-vjh9m\") pod \"nova-cell0-conductor-db-sync-wp869\" (UID: \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\") " pod="openstack/nova-cell0-conductor-db-sync-wp869" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.031398 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wp869" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.392259 4745 generic.go:334] "Generic (PLEG): container finished" podID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerID="77cfacdef3cf2a45a1b677bd0625d791d2fda0d512eff839b0ddfb40ee9d6a1c" exitCode=0 Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.392353 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e","Type":"ContainerDied","Data":"77cfacdef3cf2a45a1b677bd0625d791d2fda0d512eff839b0ddfb40ee9d6a1c"} Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.398427 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0a7607af-7365-45dc-9a8e-91719c20cdcf","Type":"ContainerStarted","Data":"57ba397484ef3d1e8915f5c16ac8c208a7bc2591b5f504dddc032b0248355a50"} Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.398474 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0a7607af-7365-45dc-9a8e-91719c20cdcf","Type":"ContainerStarted","Data":"ecd7d756e9c16aa0725e231c6eb76a740119ae14fcf145c620f089212f91500e"} Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.404234 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d2ba148a-0fdd-45d3-91e7-106c98cc7d2e","Type":"ContainerStarted","Data":"0fc92ca56759838ca327726ab014681a8459652f789825086ddd27b4f3af5054"} Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.407715 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"513c43ae-c96a-4ef0-8162-708490db52d6","Type":"ContainerStarted","Data":"860a0b8f3fffebcf82d4646e791c9d76bc4cc1eb7dc8cee74021e99f823bcbcd"} Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.440985 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.440969428 podStartE2EDuration="3.440969428s" podCreationTimestamp="2025-10-10 13:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:34:56.425360758 +0000 UTC m=+1010.323017531" watchObservedRunningTime="2025-10-10 13:34:56.440969428 +0000 UTC m=+1010.338626191" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.456554 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.456536057 podStartE2EDuration="3.456536057s" podCreationTimestamp="2025-10-10 13:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:34:56.453928182 +0000 UTC m=+1010.351584945" watchObservedRunningTime="2025-10-10 13:34:56.456536057 +0000 UTC m=+1010.354192820" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.514375 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wp869"] Oct 10 13:34:56 crc kubenswrapper[4745]: W1010 13:34:56.519749 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod907a33f4_60a8_44b7_8b1d_5fb979a84e62.slice/crio-053399218f7352066e4c0315051ab9c135b1ee570f6545b9e23ee176d81b2233 WatchSource:0}: Error finding container 053399218f7352066e4c0315051ab9c135b1ee570f6545b9e23ee176d81b2233: Status 404 returned error can't find the container with id 053399218f7352066e4c0315051ab9c135b1ee570f6545b9e23ee176d81b2233 Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.520237 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.637955 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-combined-ca-bundle\") pod \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.638027 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-run-httpd\") pod \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.638107 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gts5h\" (UniqueName: \"kubernetes.io/projected/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-kube-api-access-gts5h\") pod \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.638171 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-config-data\") pod \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.638226 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-scripts\") pod \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.638285 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-sg-core-conf-yaml\") pod \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.638309 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-log-httpd\") pod \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\" (UID: \"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e\") " Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.638495 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" (UID: "0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.638819 4745 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.639650 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" (UID: "0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.642388 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-kube-api-access-gts5h" (OuterVolumeSpecName: "kube-api-access-gts5h") pod "0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" (UID: "0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e"). InnerVolumeSpecName "kube-api-access-gts5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.652168 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-scripts" (OuterVolumeSpecName: "scripts") pod "0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" (UID: "0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.668407 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" (UID: "0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.732852 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" (UID: "0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.744005 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gts5h\" (UniqueName: \"kubernetes.io/projected/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-kube-api-access-gts5h\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.744038 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.744048 4745 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.744058 4745 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.744067 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.755722 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-config-data" (OuterVolumeSpecName: "config-data") pod "0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" (UID: "0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.837424 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 10 13:34:56 crc kubenswrapper[4745]: I1010 13:34:56.845559 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.425701 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e","Type":"ContainerDied","Data":"de8bbaebd370999f0f013c3edf3a505a21c99d41af775d4696711029257a706a"} Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.426035 4745 scope.go:117] "RemoveContainer" containerID="af89481d5714b13b02e6f904cb978eb8398047074b820c3b944fbc05ec6cf395" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.425825 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.430311 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0a7607af-7365-45dc-9a8e-91719c20cdcf","Type":"ContainerStarted","Data":"e6046aec1655bddc113c8b74b30a4f341613cbbb4f843f25f1c0b7fab41b60c5"} Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.432185 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wp869" event={"ID":"907a33f4-60a8-44b7-8b1d-5fb979a84e62","Type":"ContainerStarted","Data":"053399218f7352066e4c0315051ab9c135b1ee570f6545b9e23ee176d81b2233"} Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.458596 4745 scope.go:117] "RemoveContainer" containerID="2b54a0904603f5c65ffd8de935e25f60af0c9c3a91fae496037bcae292a76748" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.458598 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.458580819 podStartE2EDuration="3.458580819s" podCreationTimestamp="2025-10-10 13:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:34:57.447222715 +0000 UTC m=+1011.344879478" watchObservedRunningTime="2025-10-10 13:34:57.458580819 +0000 UTC m=+1011.356237582" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.480203 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.498793 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.512578 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:34:57 crc kubenswrapper[4745]: E1010 13:34:57.513885 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerName="sg-core" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.513931 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerName="sg-core" Oct 10 13:34:57 crc kubenswrapper[4745]: E1010 13:34:57.513961 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerName="proxy-httpd" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.513974 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerName="proxy-httpd" Oct 10 13:34:57 crc kubenswrapper[4745]: E1010 13:34:57.514003 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerName="ceilometer-notification-agent" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.514018 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerName="ceilometer-notification-agent" Oct 10 13:34:57 crc kubenswrapper[4745]: E1010 13:34:57.514064 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerName="ceilometer-central-agent" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.514077 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerName="ceilometer-central-agent" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.514254 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerName="proxy-httpd" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.514273 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerName="ceilometer-central-agent" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.514286 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerName="sg-core" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.514301 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" containerName="ceilometer-notification-agent" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.516184 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.521033 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.521895 4745 scope.go:117] "RemoveContainer" containerID="b7eeebdff506435ad3bd0575bea7033625902f7f515364ba9d565d65cc22cd70" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.522264 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.523071 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.563916 4745 scope.go:117] "RemoveContainer" containerID="77cfacdef3cf2a45a1b677bd0625d791d2fda0d512eff839b0ddfb40ee9d6a1c" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.668058 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fc8e852b-c861-4506-ab6f-9994a5cbe72d-log-httpd\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.668131 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-config-data\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.668260 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5p2q\" (UniqueName: \"kubernetes.io/projected/fc8e852b-c861-4506-ab6f-9994a5cbe72d-kube-api-access-c5p2q\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.668362 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fc8e852b-c861-4506-ab6f-9994a5cbe72d-run-httpd\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.668397 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-scripts\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.668626 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.668873 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.770074 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-config-data\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.770142 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5p2q\" (UniqueName: \"kubernetes.io/projected/fc8e852b-c861-4506-ab6f-9994a5cbe72d-kube-api-access-c5p2q\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.770197 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fc8e852b-c861-4506-ab6f-9994a5cbe72d-run-httpd\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.770218 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-scripts\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.770265 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.770315 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.770344 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fc8e852b-c861-4506-ab6f-9994a5cbe72d-log-httpd\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.771038 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fc8e852b-c861-4506-ab6f-9994a5cbe72d-run-httpd\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.771091 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fc8e852b-c861-4506-ab6f-9994a5cbe72d-log-httpd\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.775592 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-config-data\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.775887 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.777355 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.777782 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-scripts\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.788201 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5p2q\" (UniqueName: \"kubernetes.io/projected/fc8e852b-c861-4506-ab6f-9994a5cbe72d-kube-api-access-c5p2q\") pod \"ceilometer-0\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " pod="openstack/ceilometer-0" Oct 10 13:34:57 crc kubenswrapper[4745]: I1010 13:34:57.853299 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:34:58 crc kubenswrapper[4745]: W1010 13:34:58.291464 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc8e852b_c861_4506_ab6f_9994a5cbe72d.slice/crio-e39b428453da463ebeaae4f4185972c7b81891447f615593b03333fb0d634a99 WatchSource:0}: Error finding container e39b428453da463ebeaae4f4185972c7b81891447f615593b03333fb0d634a99: Status 404 returned error can't find the container with id e39b428453da463ebeaae4f4185972c7b81891447f615593b03333fb0d634a99 Oct 10 13:34:58 crc kubenswrapper[4745]: I1010 13:34:58.292064 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:34:58 crc kubenswrapper[4745]: I1010 13:34:58.447708 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fc8e852b-c861-4506-ab6f-9994a5cbe72d","Type":"ContainerStarted","Data":"e39b428453da463ebeaae4f4185972c7b81891447f615593b03333fb0d634a99"} Oct 10 13:34:58 crc kubenswrapper[4745]: I1010 13:34:58.755923 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e" path="/var/lib/kubelet/pods/0a2ae5f6-7659-497c-a0bd-d32e9c5cf41e/volumes" Oct 10 13:34:59 crc kubenswrapper[4745]: I1010 13:34:59.096084 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 10 13:34:59 crc kubenswrapper[4745]: I1010 13:34:59.477748 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fc8e852b-c861-4506-ab6f-9994a5cbe72d","Type":"ContainerStarted","Data":"43e933c97a8356b6a9f1f7a2b0719d9473dbb5ca5ced7154c6c3ad1e7cd4bb1d"} Oct 10 13:35:00 crc kubenswrapper[4745]: I1010 13:35:00.491252 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fc8e852b-c861-4506-ab6f-9994a5cbe72d","Type":"ContainerStarted","Data":"d281eb8b85f7e3cedef128ce8c4d7bd51eab46d48fdbe85f4086e3d41c7bd806"} Oct 10 13:35:03 crc kubenswrapper[4745]: E1010 13:35:03.502435 4745 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45009f2e_d0d5_4a20_b62c_d5d0534e1551.slice/crio-a85b3684c58b6faeef3d56c54eb1502b33619172c3103963b89eca08d70b6f4c\": RecentStats: unable to find data in memory cache]" Oct 10 13:35:03 crc kubenswrapper[4745]: I1010 13:35:03.728539 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 10 13:35:03 crc kubenswrapper[4745]: I1010 13:35:03.728585 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 10 13:35:03 crc kubenswrapper[4745]: I1010 13:35:03.766584 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 10 13:35:03 crc kubenswrapper[4745]: I1010 13:35:03.778485 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 10 13:35:04 crc kubenswrapper[4745]: I1010 13:35:04.307032 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 10 13:35:04 crc kubenswrapper[4745]: I1010 13:35:04.527238 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 10 13:35:04 crc kubenswrapper[4745]: I1010 13:35:04.527278 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 10 13:35:04 crc kubenswrapper[4745]: I1010 13:35:04.777207 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 10 13:35:04 crc kubenswrapper[4745]: I1010 13:35:04.777341 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 10 13:35:04 crc kubenswrapper[4745]: I1010 13:35:04.813914 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 10 13:35:04 crc kubenswrapper[4745]: I1010 13:35:04.822485 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 10 13:35:05 crc kubenswrapper[4745]: I1010 13:35:05.533676 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 10 13:35:05 crc kubenswrapper[4745]: I1010 13:35:05.534068 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 10 13:35:06 crc kubenswrapper[4745]: I1010 13:35:06.541823 4745 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 10 13:35:06 crc kubenswrapper[4745]: I1010 13:35:06.541874 4745 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 10 13:35:06 crc kubenswrapper[4745]: I1010 13:35:06.560377 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 10 13:35:06 crc kubenswrapper[4745]: I1010 13:35:06.561852 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 10 13:35:07 crc kubenswrapper[4745]: I1010 13:35:07.419073 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 10 13:35:07 crc kubenswrapper[4745]: I1010 13:35:07.520001 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 10 13:35:07 crc kubenswrapper[4745]: I1010 13:35:07.553519 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fc8e852b-c861-4506-ab6f-9994a5cbe72d","Type":"ContainerStarted","Data":"d342b70d70f95d528a84e1d535b35ea7acb11c89e7db5b4feca0931be5217a5c"} Oct 10 13:35:07 crc kubenswrapper[4745]: I1010 13:35:07.559289 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wp869" event={"ID":"907a33f4-60a8-44b7-8b1d-5fb979a84e62","Type":"ContainerStarted","Data":"c5fbdc3f1f579f71da481b16a039a702ef7b6c33d61cec0f03d77769c1e4cb9e"} Oct 10 13:35:07 crc kubenswrapper[4745]: I1010 13:35:07.584158 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-wp869" podStartSLOduration=1.916815981 podStartE2EDuration="12.584141282s" podCreationTimestamp="2025-10-10 13:34:55 +0000 UTC" firstStartedPulling="2025-10-10 13:34:56.529829229 +0000 UTC m=+1010.427485992" lastFinishedPulling="2025-10-10 13:35:07.19715453 +0000 UTC m=+1021.094811293" observedRunningTime="2025-10-10 13:35:07.581342502 +0000 UTC m=+1021.478999255" watchObservedRunningTime="2025-10-10 13:35:07.584141282 +0000 UTC m=+1021.481798035" Oct 10 13:35:08 crc kubenswrapper[4745]: I1010 13:35:08.884294 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:35:10 crc kubenswrapper[4745]: I1010 13:35:10.581964 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fc8e852b-c861-4506-ab6f-9994a5cbe72d","Type":"ContainerStarted","Data":"f31a93e54973a54b0849f6cdfcb1a8b5d9eaaeb94b47720c8b6a12c09a3f479e"} Oct 10 13:35:10 crc kubenswrapper[4745]: I1010 13:35:10.582655 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerName="ceilometer-central-agent" containerID="cri-o://43e933c97a8356b6a9f1f7a2b0719d9473dbb5ca5ced7154c6c3ad1e7cd4bb1d" gracePeriod=30 Oct 10 13:35:10 crc kubenswrapper[4745]: I1010 13:35:10.582709 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerName="proxy-httpd" containerID="cri-o://f31a93e54973a54b0849f6cdfcb1a8b5d9eaaeb94b47720c8b6a12c09a3f479e" gracePeriod=30 Oct 10 13:35:10 crc kubenswrapper[4745]: I1010 13:35:10.582763 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 13:35:10 crc kubenswrapper[4745]: I1010 13:35:10.582854 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerName="sg-core" containerID="cri-o://d342b70d70f95d528a84e1d535b35ea7acb11c89e7db5b4feca0931be5217a5c" gracePeriod=30 Oct 10 13:35:10 crc kubenswrapper[4745]: I1010 13:35:10.582876 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerName="ceilometer-notification-agent" containerID="cri-o://d281eb8b85f7e3cedef128ce8c4d7bd51eab46d48fdbe85f4086e3d41c7bd806" gracePeriod=30 Oct 10 13:35:10 crc kubenswrapper[4745]: I1010 13:35:10.609656 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.583684557 podStartE2EDuration="13.609641313s" podCreationTimestamp="2025-10-10 13:34:57 +0000 UTC" firstStartedPulling="2025-10-10 13:34:58.293417773 +0000 UTC m=+1012.191074536" lastFinishedPulling="2025-10-10 13:35:10.319374499 +0000 UTC m=+1024.217031292" observedRunningTime="2025-10-10 13:35:10.608237038 +0000 UTC m=+1024.505893811" watchObservedRunningTime="2025-10-10 13:35:10.609641313 +0000 UTC m=+1024.507298076" Oct 10 13:35:11 crc kubenswrapper[4745]: I1010 13:35:11.597386 4745 generic.go:334] "Generic (PLEG): container finished" podID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerID="d342b70d70f95d528a84e1d535b35ea7acb11c89e7db5b4feca0931be5217a5c" exitCode=2 Oct 10 13:35:11 crc kubenswrapper[4745]: I1010 13:35:11.597787 4745 generic.go:334] "Generic (PLEG): container finished" podID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerID="d281eb8b85f7e3cedef128ce8c4d7bd51eab46d48fdbe85f4086e3d41c7bd806" exitCode=0 Oct 10 13:35:11 crc kubenswrapper[4745]: I1010 13:35:11.597800 4745 generic.go:334] "Generic (PLEG): container finished" podID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerID="43e933c97a8356b6a9f1f7a2b0719d9473dbb5ca5ced7154c6c3ad1e7cd4bb1d" exitCode=0 Oct 10 13:35:11 crc kubenswrapper[4745]: I1010 13:35:11.597422 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fc8e852b-c861-4506-ab6f-9994a5cbe72d","Type":"ContainerDied","Data":"d342b70d70f95d528a84e1d535b35ea7acb11c89e7db5b4feca0931be5217a5c"} Oct 10 13:35:11 crc kubenswrapper[4745]: I1010 13:35:11.597845 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fc8e852b-c861-4506-ab6f-9994a5cbe72d","Type":"ContainerDied","Data":"d281eb8b85f7e3cedef128ce8c4d7bd51eab46d48fdbe85f4086e3d41c7bd806"} Oct 10 13:35:11 crc kubenswrapper[4745]: I1010 13:35:11.597860 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fc8e852b-c861-4506-ab6f-9994a5cbe72d","Type":"ContainerDied","Data":"43e933c97a8356b6a9f1f7a2b0719d9473dbb5ca5ced7154c6c3ad1e7cd4bb1d"} Oct 10 13:35:18 crc kubenswrapper[4745]: I1010 13:35:18.677018 4745 generic.go:334] "Generic (PLEG): container finished" podID="907a33f4-60a8-44b7-8b1d-5fb979a84e62" containerID="c5fbdc3f1f579f71da481b16a039a702ef7b6c33d61cec0f03d77769c1e4cb9e" exitCode=0 Oct 10 13:35:18 crc kubenswrapper[4745]: I1010 13:35:18.677104 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wp869" event={"ID":"907a33f4-60a8-44b7-8b1d-5fb979a84e62","Type":"ContainerDied","Data":"c5fbdc3f1f579f71da481b16a039a702ef7b6c33d61cec0f03d77769c1e4cb9e"} Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.161319 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wp869" Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.312622 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-config-data\") pod \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\" (UID: \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\") " Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.312673 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-scripts\") pod \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\" (UID: \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\") " Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.312747 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-combined-ca-bundle\") pod \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\" (UID: \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\") " Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.312771 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjh9m\" (UniqueName: \"kubernetes.io/projected/907a33f4-60a8-44b7-8b1d-5fb979a84e62-kube-api-access-vjh9m\") pod \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\" (UID: \"907a33f4-60a8-44b7-8b1d-5fb979a84e62\") " Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.319077 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/907a33f4-60a8-44b7-8b1d-5fb979a84e62-kube-api-access-vjh9m" (OuterVolumeSpecName: "kube-api-access-vjh9m") pod "907a33f4-60a8-44b7-8b1d-5fb979a84e62" (UID: "907a33f4-60a8-44b7-8b1d-5fb979a84e62"). InnerVolumeSpecName "kube-api-access-vjh9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.319171 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-scripts" (OuterVolumeSpecName: "scripts") pod "907a33f4-60a8-44b7-8b1d-5fb979a84e62" (UID: "907a33f4-60a8-44b7-8b1d-5fb979a84e62"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.354383 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-config-data" (OuterVolumeSpecName: "config-data") pod "907a33f4-60a8-44b7-8b1d-5fb979a84e62" (UID: "907a33f4-60a8-44b7-8b1d-5fb979a84e62"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.361058 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "907a33f4-60a8-44b7-8b1d-5fb979a84e62" (UID: "907a33f4-60a8-44b7-8b1d-5fb979a84e62"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.415553 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjh9m\" (UniqueName: \"kubernetes.io/projected/907a33f4-60a8-44b7-8b1d-5fb979a84e62-kube-api-access-vjh9m\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.415654 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.415675 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.415695 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/907a33f4-60a8-44b7-8b1d-5fb979a84e62-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.705962 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wp869" event={"ID":"907a33f4-60a8-44b7-8b1d-5fb979a84e62","Type":"ContainerDied","Data":"053399218f7352066e4c0315051ab9c135b1ee570f6545b9e23ee176d81b2233"} Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.706324 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="053399218f7352066e4c0315051ab9c135b1ee570f6545b9e23ee176d81b2233" Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.706054 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wp869" Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.880181 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 13:35:20 crc kubenswrapper[4745]: E1010 13:35:20.881154 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="907a33f4-60a8-44b7-8b1d-5fb979a84e62" containerName="nova-cell0-conductor-db-sync" Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.881328 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="907a33f4-60a8-44b7-8b1d-5fb979a84e62" containerName="nova-cell0-conductor-db-sync" Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.881867 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="907a33f4-60a8-44b7-8b1d-5fb979a84e62" containerName="nova-cell0-conductor-db-sync" Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.883102 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.888346 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.888612 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bx7wb" Oct 10 13:35:20 crc kubenswrapper[4745]: I1010 13:35:20.891564 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 13:35:21 crc kubenswrapper[4745]: I1010 13:35:21.026991 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac183214-b5c6-4e33-8045-ddef84c43323-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ac183214-b5c6-4e33-8045-ddef84c43323\") " pod="openstack/nova-cell0-conductor-0" Oct 10 13:35:21 crc kubenswrapper[4745]: I1010 13:35:21.027346 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrjqw\" (UniqueName: \"kubernetes.io/projected/ac183214-b5c6-4e33-8045-ddef84c43323-kube-api-access-vrjqw\") pod \"nova-cell0-conductor-0\" (UID: \"ac183214-b5c6-4e33-8045-ddef84c43323\") " pod="openstack/nova-cell0-conductor-0" Oct 10 13:35:21 crc kubenswrapper[4745]: I1010 13:35:21.027488 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac183214-b5c6-4e33-8045-ddef84c43323-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ac183214-b5c6-4e33-8045-ddef84c43323\") " pod="openstack/nova-cell0-conductor-0" Oct 10 13:35:21 crc kubenswrapper[4745]: I1010 13:35:21.129136 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac183214-b5c6-4e33-8045-ddef84c43323-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ac183214-b5c6-4e33-8045-ddef84c43323\") " pod="openstack/nova-cell0-conductor-0" Oct 10 13:35:21 crc kubenswrapper[4745]: I1010 13:35:21.129278 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrjqw\" (UniqueName: \"kubernetes.io/projected/ac183214-b5c6-4e33-8045-ddef84c43323-kube-api-access-vrjqw\") pod \"nova-cell0-conductor-0\" (UID: \"ac183214-b5c6-4e33-8045-ddef84c43323\") " pod="openstack/nova-cell0-conductor-0" Oct 10 13:35:21 crc kubenswrapper[4745]: I1010 13:35:21.129310 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac183214-b5c6-4e33-8045-ddef84c43323-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ac183214-b5c6-4e33-8045-ddef84c43323\") " pod="openstack/nova-cell0-conductor-0" Oct 10 13:35:21 crc kubenswrapper[4745]: I1010 13:35:21.133395 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac183214-b5c6-4e33-8045-ddef84c43323-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ac183214-b5c6-4e33-8045-ddef84c43323\") " pod="openstack/nova-cell0-conductor-0" Oct 10 13:35:21 crc kubenswrapper[4745]: I1010 13:35:21.134485 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac183214-b5c6-4e33-8045-ddef84c43323-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ac183214-b5c6-4e33-8045-ddef84c43323\") " pod="openstack/nova-cell0-conductor-0" Oct 10 13:35:21 crc kubenswrapper[4745]: I1010 13:35:21.157091 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrjqw\" (UniqueName: \"kubernetes.io/projected/ac183214-b5c6-4e33-8045-ddef84c43323-kube-api-access-vrjqw\") pod \"nova-cell0-conductor-0\" (UID: \"ac183214-b5c6-4e33-8045-ddef84c43323\") " pod="openstack/nova-cell0-conductor-0" Oct 10 13:35:21 crc kubenswrapper[4745]: I1010 13:35:21.248762 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 13:35:21 crc kubenswrapper[4745]: I1010 13:35:21.718468 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 13:35:22 crc kubenswrapper[4745]: I1010 13:35:22.725653 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ac183214-b5c6-4e33-8045-ddef84c43323","Type":"ContainerStarted","Data":"92403e62c0ae52c1672ce616252b79e5c5dcca83d9e4c385ff910e666385381a"} Oct 10 13:35:22 crc kubenswrapper[4745]: I1010 13:35:22.726062 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 10 13:35:22 crc kubenswrapper[4745]: I1010 13:35:22.726079 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ac183214-b5c6-4e33-8045-ddef84c43323","Type":"ContainerStarted","Data":"39245079834e7e9f0ecc1a15a946ea62aa3af15786197eab55675e76d38a8215"} Oct 10 13:35:22 crc kubenswrapper[4745]: I1010 13:35:22.763296 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.763268597 podStartE2EDuration="2.763268597s" podCreationTimestamp="2025-10-10 13:35:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:35:22.748777225 +0000 UTC m=+1036.646434008" watchObservedRunningTime="2025-10-10 13:35:22.763268597 +0000 UTC m=+1036.660925400" Oct 10 13:35:26 crc kubenswrapper[4745]: I1010 13:35:26.284001 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 10 13:35:26 crc kubenswrapper[4745]: I1010 13:35:26.879852 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-8t6r5"] Oct 10 13:35:26 crc kubenswrapper[4745]: I1010 13:35:26.881181 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8t6r5" Oct 10 13:35:26 crc kubenswrapper[4745]: I1010 13:35:26.884227 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 10 13:35:26 crc kubenswrapper[4745]: I1010 13:35:26.884407 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 10 13:35:26 crc kubenswrapper[4745]: I1010 13:35:26.893569 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-8t6r5"] Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.054619 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-config-data\") pod \"nova-cell0-cell-mapping-8t6r5\" (UID: \"50933493-e7d7-4fc2-93bc-69173739fe58\") " pod="openstack/nova-cell0-cell-mapping-8t6r5" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.054971 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-scripts\") pod \"nova-cell0-cell-mapping-8t6r5\" (UID: \"50933493-e7d7-4fc2-93bc-69173739fe58\") " pod="openstack/nova-cell0-cell-mapping-8t6r5" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.054995 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcp7r\" (UniqueName: \"kubernetes.io/projected/50933493-e7d7-4fc2-93bc-69173739fe58-kube-api-access-kcp7r\") pod \"nova-cell0-cell-mapping-8t6r5\" (UID: \"50933493-e7d7-4fc2-93bc-69173739fe58\") " pod="openstack/nova-cell0-cell-mapping-8t6r5" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.055054 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-8t6r5\" (UID: \"50933493-e7d7-4fc2-93bc-69173739fe58\") " pod="openstack/nova-cell0-cell-mapping-8t6r5" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.057676 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.059027 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.074182 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.080208 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.138624 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.141810 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.145706 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.157139 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a65d55af-b9a5-4f6f-b62d-5683e387caa5-config-data\") pod \"nova-scheduler-0\" (UID: \"a65d55af-b9a5-4f6f-b62d-5683e387caa5\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.157196 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-8t6r5\" (UID: \"50933493-e7d7-4fc2-93bc-69173739fe58\") " pod="openstack/nova-cell0-cell-mapping-8t6r5" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.157239 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-config-data\") pod \"nova-cell0-cell-mapping-8t6r5\" (UID: \"50933493-e7d7-4fc2-93bc-69173739fe58\") " pod="openstack/nova-cell0-cell-mapping-8t6r5" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.157291 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk8mr\" (UniqueName: \"kubernetes.io/projected/a65d55af-b9a5-4f6f-b62d-5683e387caa5-kube-api-access-tk8mr\") pod \"nova-scheduler-0\" (UID: \"a65d55af-b9a5-4f6f-b62d-5683e387caa5\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.157316 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a65d55af-b9a5-4f6f-b62d-5683e387caa5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a65d55af-b9a5-4f6f-b62d-5683e387caa5\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.157351 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-scripts\") pod \"nova-cell0-cell-mapping-8t6r5\" (UID: \"50933493-e7d7-4fc2-93bc-69173739fe58\") " pod="openstack/nova-cell0-cell-mapping-8t6r5" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.157372 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcp7r\" (UniqueName: \"kubernetes.io/projected/50933493-e7d7-4fc2-93bc-69173739fe58-kube-api-access-kcp7r\") pod \"nova-cell0-cell-mapping-8t6r5\" (UID: \"50933493-e7d7-4fc2-93bc-69173739fe58\") " pod="openstack/nova-cell0-cell-mapping-8t6r5" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.168487 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-config-data\") pod \"nova-cell0-cell-mapping-8t6r5\" (UID: \"50933493-e7d7-4fc2-93bc-69173739fe58\") " pod="openstack/nova-cell0-cell-mapping-8t6r5" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.185332 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-8t6r5\" (UID: \"50933493-e7d7-4fc2-93bc-69173739fe58\") " pod="openstack/nova-cell0-cell-mapping-8t6r5" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.191493 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-scripts\") pod \"nova-cell0-cell-mapping-8t6r5\" (UID: \"50933493-e7d7-4fc2-93bc-69173739fe58\") " pod="openstack/nova-cell0-cell-mapping-8t6r5" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.198605 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.205227 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcp7r\" (UniqueName: \"kubernetes.io/projected/50933493-e7d7-4fc2-93bc-69173739fe58-kube-api-access-kcp7r\") pod \"nova-cell0-cell-mapping-8t6r5\" (UID: \"50933493-e7d7-4fc2-93bc-69173739fe58\") " pod="openstack/nova-cell0-cell-mapping-8t6r5" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.215105 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8t6r5" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.215584 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.217046 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.221768 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.234883 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.236479 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.244680 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.246746 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.259316 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a65d55af-b9a5-4f6f-b62d-5683e387caa5-config-data\") pod \"nova-scheduler-0\" (UID: \"a65d55af-b9a5-4f6f-b62d-5683e387caa5\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.259435 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk8mr\" (UniqueName: \"kubernetes.io/projected/a65d55af-b9a5-4f6f-b62d-5683e387caa5-kube-api-access-tk8mr\") pod \"nova-scheduler-0\" (UID: \"a65d55af-b9a5-4f6f-b62d-5683e387caa5\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.259466 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a65d55af-b9a5-4f6f-b62d-5683e387caa5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a65d55af-b9a5-4f6f-b62d-5683e387caa5\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.259504 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frt9x\" (UniqueName: \"kubernetes.io/projected/24ee955e-b7b3-43c2-8dbe-7118951f037d-kube-api-access-frt9x\") pod \"nova-api-0\" (UID: \"24ee955e-b7b3-43c2-8dbe-7118951f037d\") " pod="openstack/nova-api-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.259526 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ee955e-b7b3-43c2-8dbe-7118951f037d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"24ee955e-b7b3-43c2-8dbe-7118951f037d\") " pod="openstack/nova-api-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.259556 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ee955e-b7b3-43c2-8dbe-7118951f037d-config-data\") pod \"nova-api-0\" (UID: \"24ee955e-b7b3-43c2-8dbe-7118951f037d\") " pod="openstack/nova-api-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.259580 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24ee955e-b7b3-43c2-8dbe-7118951f037d-logs\") pod \"nova-api-0\" (UID: \"24ee955e-b7b3-43c2-8dbe-7118951f037d\") " pod="openstack/nova-api-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.271398 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a65d55af-b9a5-4f6f-b62d-5683e387caa5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a65d55af-b9a5-4f6f-b62d-5683e387caa5\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.275974 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.307219 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a65d55af-b9a5-4f6f-b62d-5683e387caa5-config-data\") pod \"nova-scheduler-0\" (UID: \"a65d55af-b9a5-4f6f-b62d-5683e387caa5\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.311274 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk8mr\" (UniqueName: \"kubernetes.io/projected/a65d55af-b9a5-4f6f-b62d-5683e387caa5-kube-api-access-tk8mr\") pod \"nova-scheduler-0\" (UID: \"a65d55af-b9a5-4f6f-b62d-5683e387caa5\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.323021 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-cp2b9"] Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.324714 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.359116 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-cp2b9"] Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.360658 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77wps\" (UniqueName: \"kubernetes.io/projected/89b5e6b9-5dac-4856-83b3-fb61388231dc-kube-api-access-77wps\") pod \"nova-cell1-novncproxy-0\" (UID: \"89b5e6b9-5dac-4856-83b3-fb61388231dc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.360703 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frt9x\" (UniqueName: \"kubernetes.io/projected/24ee955e-b7b3-43c2-8dbe-7118951f037d-kube-api-access-frt9x\") pod \"nova-api-0\" (UID: \"24ee955e-b7b3-43c2-8dbe-7118951f037d\") " pod="openstack/nova-api-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.360740 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ee955e-b7b3-43c2-8dbe-7118951f037d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"24ee955e-b7b3-43c2-8dbe-7118951f037d\") " pod="openstack/nova-api-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.360769 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ee955e-b7b3-43c2-8dbe-7118951f037d-config-data\") pod \"nova-api-0\" (UID: \"24ee955e-b7b3-43c2-8dbe-7118951f037d\") " pod="openstack/nova-api-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.360793 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24ee955e-b7b3-43c2-8dbe-7118951f037d-logs\") pod \"nova-api-0\" (UID: \"24ee955e-b7b3-43c2-8dbe-7118951f037d\") " pod="openstack/nova-api-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.360810 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b434a74-d01a-4af8-8d80-facf27b95128-config-data\") pod \"nova-metadata-0\" (UID: \"5b434a74-d01a-4af8-8d80-facf27b95128\") " pod="openstack/nova-metadata-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.360838 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vxjs\" (UniqueName: \"kubernetes.io/projected/5b434a74-d01a-4af8-8d80-facf27b95128-kube-api-access-5vxjs\") pod \"nova-metadata-0\" (UID: \"5b434a74-d01a-4af8-8d80-facf27b95128\") " pod="openstack/nova-metadata-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.360927 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b434a74-d01a-4af8-8d80-facf27b95128-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5b434a74-d01a-4af8-8d80-facf27b95128\") " pod="openstack/nova-metadata-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.360973 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b5e6b9-5dac-4856-83b3-fb61388231dc-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"89b5e6b9-5dac-4856-83b3-fb61388231dc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.360997 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b434a74-d01a-4af8-8d80-facf27b95128-logs\") pod \"nova-metadata-0\" (UID: \"5b434a74-d01a-4af8-8d80-facf27b95128\") " pod="openstack/nova-metadata-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.361017 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b5e6b9-5dac-4856-83b3-fb61388231dc-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"89b5e6b9-5dac-4856-83b3-fb61388231dc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.361398 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24ee955e-b7b3-43c2-8dbe-7118951f037d-logs\") pod \"nova-api-0\" (UID: \"24ee955e-b7b3-43c2-8dbe-7118951f037d\") " pod="openstack/nova-api-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.365447 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ee955e-b7b3-43c2-8dbe-7118951f037d-config-data\") pod \"nova-api-0\" (UID: \"24ee955e-b7b3-43c2-8dbe-7118951f037d\") " pod="openstack/nova-api-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.365765 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ee955e-b7b3-43c2-8dbe-7118951f037d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"24ee955e-b7b3-43c2-8dbe-7118951f037d\") " pod="openstack/nova-api-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.373046 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.389517 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frt9x\" (UniqueName: \"kubernetes.io/projected/24ee955e-b7b3-43c2-8dbe-7118951f037d-kube-api-access-frt9x\") pod \"nova-api-0\" (UID: \"24ee955e-b7b3-43c2-8dbe-7118951f037d\") " pod="openstack/nova-api-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.462533 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77wps\" (UniqueName: \"kubernetes.io/projected/89b5e6b9-5dac-4856-83b3-fb61388231dc-kube-api-access-77wps\") pod \"nova-cell1-novncproxy-0\" (UID: \"89b5e6b9-5dac-4856-83b3-fb61388231dc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.462580 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srb4n\" (UniqueName: \"kubernetes.io/projected/ab0f388b-1766-4cf5-8d43-6811ab7263d5-kube-api-access-srb4n\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.462612 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.462647 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b434a74-d01a-4af8-8d80-facf27b95128-config-data\") pod \"nova-metadata-0\" (UID: \"5b434a74-d01a-4af8-8d80-facf27b95128\") " pod="openstack/nova-metadata-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.462663 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.462687 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vxjs\" (UniqueName: \"kubernetes.io/projected/5b434a74-d01a-4af8-8d80-facf27b95128-kube-api-access-5vxjs\") pod \"nova-metadata-0\" (UID: \"5b434a74-d01a-4af8-8d80-facf27b95128\") " pod="openstack/nova-metadata-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.462715 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.462814 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b434a74-d01a-4af8-8d80-facf27b95128-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5b434a74-d01a-4af8-8d80-facf27b95128\") " pod="openstack/nova-metadata-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.462837 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b5e6b9-5dac-4856-83b3-fb61388231dc-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"89b5e6b9-5dac-4856-83b3-fb61388231dc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.462858 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b434a74-d01a-4af8-8d80-facf27b95128-logs\") pod \"nova-metadata-0\" (UID: \"5b434a74-d01a-4af8-8d80-facf27b95128\") " pod="openstack/nova-metadata-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.462879 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b5e6b9-5dac-4856-83b3-fb61388231dc-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"89b5e6b9-5dac-4856-83b3-fb61388231dc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.462896 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.462938 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-config\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.465084 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b434a74-d01a-4af8-8d80-facf27b95128-logs\") pod \"nova-metadata-0\" (UID: \"5b434a74-d01a-4af8-8d80-facf27b95128\") " pod="openstack/nova-metadata-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.465474 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.471225 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b434a74-d01a-4af8-8d80-facf27b95128-config-data\") pod \"nova-metadata-0\" (UID: \"5b434a74-d01a-4af8-8d80-facf27b95128\") " pod="openstack/nova-metadata-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.473249 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b5e6b9-5dac-4856-83b3-fb61388231dc-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"89b5e6b9-5dac-4856-83b3-fb61388231dc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.473391 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b5e6b9-5dac-4856-83b3-fb61388231dc-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"89b5e6b9-5dac-4856-83b3-fb61388231dc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.486952 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b434a74-d01a-4af8-8d80-facf27b95128-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5b434a74-d01a-4af8-8d80-facf27b95128\") " pod="openstack/nova-metadata-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.487654 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vxjs\" (UniqueName: \"kubernetes.io/projected/5b434a74-d01a-4af8-8d80-facf27b95128-kube-api-access-5vxjs\") pod \"nova-metadata-0\" (UID: \"5b434a74-d01a-4af8-8d80-facf27b95128\") " pod="openstack/nova-metadata-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.491428 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77wps\" (UniqueName: \"kubernetes.io/projected/89b5e6b9-5dac-4856-83b3-fb61388231dc-kube-api-access-77wps\") pod \"nova-cell1-novncproxy-0\" (UID: \"89b5e6b9-5dac-4856-83b3-fb61388231dc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.564295 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srb4n\" (UniqueName: \"kubernetes.io/projected/ab0f388b-1766-4cf5-8d43-6811ab7263d5-kube-api-access-srb4n\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.564603 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.564646 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.564690 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.564772 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.564819 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-config\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.565625 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-config\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.565987 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.566183 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.566872 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.567390 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.586581 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srb4n\" (UniqueName: \"kubernetes.io/projected/ab0f388b-1766-4cf5-8d43-6811ab7263d5-kube-api-access-srb4n\") pod \"dnsmasq-dns-845d6d6f59-cp2b9\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.744299 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.769918 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.792716 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.875961 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-8t6r5"] Oct 10 13:35:27 crc kubenswrapper[4745]: I1010 13:35:27.942107 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.000056 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.005685 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 13:35:28 crc kubenswrapper[4745]: W1010 13:35:28.017548 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24ee955e_b7b3_43c2_8dbe_7118951f037d.slice/crio-fe8feed6bed477636b2223879a7b8a40b77dba076a92233e691d67d7adbd6374 WatchSource:0}: Error finding container fe8feed6bed477636b2223879a7b8a40b77dba076a92233e691d67d7adbd6374: Status 404 returned error can't find the container with id fe8feed6bed477636b2223879a7b8a40b77dba076a92233e691d67d7adbd6374 Oct 10 13:35:28 crc kubenswrapper[4745]: W1010 13:35:28.024694 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda65d55af_b9a5_4f6f_b62d_5683e387caa5.slice/crio-412d12e8fbc5c4c9ddd56c583e52695ad3a32af4cae8de8236336347dcc83dbc WatchSource:0}: Error finding container 412d12e8fbc5c4c9ddd56c583e52695ad3a32af4cae8de8236336347dcc83dbc: Status 404 returned error can't find the container with id 412d12e8fbc5c4c9ddd56c583e52695ad3a32af4cae8de8236336347dcc83dbc Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.071354 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6xlzw"] Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.073190 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6xlzw" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.076564 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.076842 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.094851 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6xlzw"] Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.209767 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-scripts\") pod \"nova-cell1-conductor-db-sync-6xlzw\" (UID: \"642aa23f-361f-4bc5-b463-15a104a8d7d7\") " pod="openstack/nova-cell1-conductor-db-sync-6xlzw" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.210170 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drtjp\" (UniqueName: \"kubernetes.io/projected/642aa23f-361f-4bc5-b463-15a104a8d7d7-kube-api-access-drtjp\") pod \"nova-cell1-conductor-db-sync-6xlzw\" (UID: \"642aa23f-361f-4bc5-b463-15a104a8d7d7\") " pod="openstack/nova-cell1-conductor-db-sync-6xlzw" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.210200 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6xlzw\" (UID: \"642aa23f-361f-4bc5-b463-15a104a8d7d7\") " pod="openstack/nova-cell1-conductor-db-sync-6xlzw" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.210240 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-config-data\") pod \"nova-cell1-conductor-db-sync-6xlzw\" (UID: \"642aa23f-361f-4bc5-b463-15a104a8d7d7\") " pod="openstack/nova-cell1-conductor-db-sync-6xlzw" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.273599 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.311756 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-config-data\") pod \"nova-cell1-conductor-db-sync-6xlzw\" (UID: \"642aa23f-361f-4bc5-b463-15a104a8d7d7\") " pod="openstack/nova-cell1-conductor-db-sync-6xlzw" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.311851 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-scripts\") pod \"nova-cell1-conductor-db-sync-6xlzw\" (UID: \"642aa23f-361f-4bc5-b463-15a104a8d7d7\") " pod="openstack/nova-cell1-conductor-db-sync-6xlzw" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.311933 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drtjp\" (UniqueName: \"kubernetes.io/projected/642aa23f-361f-4bc5-b463-15a104a8d7d7-kube-api-access-drtjp\") pod \"nova-cell1-conductor-db-sync-6xlzw\" (UID: \"642aa23f-361f-4bc5-b463-15a104a8d7d7\") " pod="openstack/nova-cell1-conductor-db-sync-6xlzw" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.311958 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6xlzw\" (UID: \"642aa23f-361f-4bc5-b463-15a104a8d7d7\") " pod="openstack/nova-cell1-conductor-db-sync-6xlzw" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.317670 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6xlzw\" (UID: \"642aa23f-361f-4bc5-b463-15a104a8d7d7\") " pod="openstack/nova-cell1-conductor-db-sync-6xlzw" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.323109 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-scripts\") pod \"nova-cell1-conductor-db-sync-6xlzw\" (UID: \"642aa23f-361f-4bc5-b463-15a104a8d7d7\") " pod="openstack/nova-cell1-conductor-db-sync-6xlzw" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.323806 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-config-data\") pod \"nova-cell1-conductor-db-sync-6xlzw\" (UID: \"642aa23f-361f-4bc5-b463-15a104a8d7d7\") " pod="openstack/nova-cell1-conductor-db-sync-6xlzw" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.336430 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drtjp\" (UniqueName: \"kubernetes.io/projected/642aa23f-361f-4bc5-b463-15a104a8d7d7-kube-api-access-drtjp\") pod \"nova-cell1-conductor-db-sync-6xlzw\" (UID: \"642aa23f-361f-4bc5-b463-15a104a8d7d7\") " pod="openstack/nova-cell1-conductor-db-sync-6xlzw" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.376818 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.402000 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-cp2b9"] Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.418963 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6xlzw" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.819932 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"89b5e6b9-5dac-4856-83b3-fb61388231dc","Type":"ContainerStarted","Data":"a93cbae6081756558d97d56f5bc5eba4354a318671727c31cbaaca4575dfd3cc"} Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.823345 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a65d55af-b9a5-4f6f-b62d-5683e387caa5","Type":"ContainerStarted","Data":"412d12e8fbc5c4c9ddd56c583e52695ad3a32af4cae8de8236336347dcc83dbc"} Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.825191 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8t6r5" event={"ID":"50933493-e7d7-4fc2-93bc-69173739fe58","Type":"ContainerStarted","Data":"41e66cac9137afc38339955332ce978e26364120893d9f3b693a00b8843750a2"} Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.825238 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8t6r5" event={"ID":"50933493-e7d7-4fc2-93bc-69173739fe58","Type":"ContainerStarted","Data":"8ed91649e2c3326d702d4f20bf49662d2d6be31f8eff598ea62d90e8b2c4b687"} Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.828864 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b434a74-d01a-4af8-8d80-facf27b95128","Type":"ContainerStarted","Data":"68955bcdad9c5815104705d1cfbe83f6a47f577350468b0cb7ca5b66cd8eeba3"} Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.849629 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-8t6r5" podStartSLOduration=2.849612924 podStartE2EDuration="2.849612924s" podCreationTimestamp="2025-10-10 13:35:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:35:28.844557398 +0000 UTC m=+1042.742214161" watchObservedRunningTime="2025-10-10 13:35:28.849612924 +0000 UTC m=+1042.747269687" Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.865580 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" event={"ID":"ab0f388b-1766-4cf5-8d43-6811ab7263d5","Type":"ContainerStarted","Data":"5c29b77ee52a4466794a32b6d88addde506044a430165b2e9100f0de056c466b"} Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.865625 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" event={"ID":"ab0f388b-1766-4cf5-8d43-6811ab7263d5","Type":"ContainerStarted","Data":"ce5fc0dbfe107bdb6a9d61cf471c35aa4ded16a847cb27d12ab24a02a595c815"} Oct 10 13:35:28 crc kubenswrapper[4745]: I1010 13:35:28.871629 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"24ee955e-b7b3-43c2-8dbe-7118951f037d","Type":"ContainerStarted","Data":"fe8feed6bed477636b2223879a7b8a40b77dba076a92233e691d67d7adbd6374"} Oct 10 13:35:29 crc kubenswrapper[4745]: I1010 13:35:29.039786 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6xlzw"] Oct 10 13:35:29 crc kubenswrapper[4745]: W1010 13:35:29.043881 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod642aa23f_361f_4bc5_b463_15a104a8d7d7.slice/crio-e121e21ce5ead97ee421c2f9f8fc096225351410b7a910208ed835180c9a8fd4 WatchSource:0}: Error finding container e121e21ce5ead97ee421c2f9f8fc096225351410b7a910208ed835180c9a8fd4: Status 404 returned error can't find the container with id e121e21ce5ead97ee421c2f9f8fc096225351410b7a910208ed835180c9a8fd4 Oct 10 13:35:29 crc kubenswrapper[4745]: I1010 13:35:29.882431 4745 generic.go:334] "Generic (PLEG): container finished" podID="ab0f388b-1766-4cf5-8d43-6811ab7263d5" containerID="5c29b77ee52a4466794a32b6d88addde506044a430165b2e9100f0de056c466b" exitCode=0 Oct 10 13:35:29 crc kubenswrapper[4745]: I1010 13:35:29.882546 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" event={"ID":"ab0f388b-1766-4cf5-8d43-6811ab7263d5","Type":"ContainerDied","Data":"5c29b77ee52a4466794a32b6d88addde506044a430165b2e9100f0de056c466b"} Oct 10 13:35:29 crc kubenswrapper[4745]: I1010 13:35:29.888275 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6xlzw" event={"ID":"642aa23f-361f-4bc5-b463-15a104a8d7d7","Type":"ContainerStarted","Data":"7852ea056a15f2decd43c0c52a5fb22a6679a34eb1eb98afd32eba762c9edd28"} Oct 10 13:35:29 crc kubenswrapper[4745]: I1010 13:35:29.888316 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6xlzw" event={"ID":"642aa23f-361f-4bc5-b463-15a104a8d7d7","Type":"ContainerStarted","Data":"e121e21ce5ead97ee421c2f9f8fc096225351410b7a910208ed835180c9a8fd4"} Oct 10 13:35:29 crc kubenswrapper[4745]: I1010 13:35:29.928776 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-6xlzw" podStartSLOduration=1.928708902 podStartE2EDuration="1.928708902s" podCreationTimestamp="2025-10-10 13:35:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:35:29.914943068 +0000 UTC m=+1043.812599831" watchObservedRunningTime="2025-10-10 13:35:29.928708902 +0000 UTC m=+1043.826365675" Oct 10 13:35:30 crc kubenswrapper[4745]: I1010 13:35:30.535961 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 13:35:30 crc kubenswrapper[4745]: I1010 13:35:30.549998 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:35:31 crc kubenswrapper[4745]: I1010 13:35:31.920859 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a65d55af-b9a5-4f6f-b62d-5683e387caa5","Type":"ContainerStarted","Data":"f673012d6f3cc5c2f4698094b96935ef912bb8c8392c8add4b3f559c0e551756"} Oct 10 13:35:31 crc kubenswrapper[4745]: I1010 13:35:31.924302 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b434a74-d01a-4af8-8d80-facf27b95128","Type":"ContainerStarted","Data":"33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead"} Oct 10 13:35:31 crc kubenswrapper[4745]: I1010 13:35:31.926500 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" event={"ID":"ab0f388b-1766-4cf5-8d43-6811ab7263d5","Type":"ContainerStarted","Data":"c5828bbc40ad4151ca605349b1eaffcd0f4ad4f1db8fd07fcee8f418d0f5e6e0"} Oct 10 13:35:31 crc kubenswrapper[4745]: I1010 13:35:31.926623 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:31 crc kubenswrapper[4745]: I1010 13:35:31.928889 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"24ee955e-b7b3-43c2-8dbe-7118951f037d","Type":"ContainerStarted","Data":"3e8c68e1a2099ddee002f06ea9351311f067ec5a8cfa088f36d3254930ad7d3b"} Oct 10 13:35:31 crc kubenswrapper[4745]: I1010 13:35:31.928932 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"24ee955e-b7b3-43c2-8dbe-7118951f037d","Type":"ContainerStarted","Data":"167f6a8d5aade40bd04faed225146f2986dc93ee34716a8fc5c88b1599d6d406"} Oct 10 13:35:31 crc kubenswrapper[4745]: I1010 13:35:31.932312 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"89b5e6b9-5dac-4856-83b3-fb61388231dc","Type":"ContainerStarted","Data":"3ec82c7f9d4232b42f193eea107a76ff7031e4f6c76dbcddbd6f83b0a7679d47"} Oct 10 13:35:31 crc kubenswrapper[4745]: I1010 13:35:31.932417 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="89b5e6b9-5dac-4856-83b3-fb61388231dc" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://3ec82c7f9d4232b42f193eea107a76ff7031e4f6c76dbcddbd6f83b0a7679d47" gracePeriod=30 Oct 10 13:35:31 crc kubenswrapper[4745]: I1010 13:35:31.939418 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.7456496860000001 podStartE2EDuration="4.939397662s" podCreationTimestamp="2025-10-10 13:35:27 +0000 UTC" firstStartedPulling="2025-10-10 13:35:28.02756795 +0000 UTC m=+1041.925224713" lastFinishedPulling="2025-10-10 13:35:31.221315926 +0000 UTC m=+1045.118972689" observedRunningTime="2025-10-10 13:35:31.935395492 +0000 UTC m=+1045.833052255" watchObservedRunningTime="2025-10-10 13:35:31.939397662 +0000 UTC m=+1045.837054435" Oct 10 13:35:31 crc kubenswrapper[4745]: I1010 13:35:31.971709 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.128384261 podStartE2EDuration="4.971689409s" podCreationTimestamp="2025-10-10 13:35:27 +0000 UTC" firstStartedPulling="2025-10-10 13:35:28.383364722 +0000 UTC m=+1042.281021485" lastFinishedPulling="2025-10-10 13:35:31.22666985 +0000 UTC m=+1045.124326633" observedRunningTime="2025-10-10 13:35:31.966905419 +0000 UTC m=+1045.864562192" watchObservedRunningTime="2025-10-10 13:35:31.971689409 +0000 UTC m=+1045.869346172" Oct 10 13:35:31 crc kubenswrapper[4745]: I1010 13:35:31.992876 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" podStartSLOduration=4.9928510379999995 podStartE2EDuration="4.992851038s" podCreationTimestamp="2025-10-10 13:35:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:35:31.987159496 +0000 UTC m=+1045.884816269" watchObservedRunningTime="2025-10-10 13:35:31.992851038 +0000 UTC m=+1045.890507801" Oct 10 13:35:32 crc kubenswrapper[4745]: I1010 13:35:32.015540 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.815321037 podStartE2EDuration="5.015524315s" podCreationTimestamp="2025-10-10 13:35:27 +0000 UTC" firstStartedPulling="2025-10-10 13:35:28.020854022 +0000 UTC m=+1041.918510785" lastFinishedPulling="2025-10-10 13:35:31.2210573 +0000 UTC m=+1045.118714063" observedRunningTime="2025-10-10 13:35:32.012106459 +0000 UTC m=+1045.909763222" watchObservedRunningTime="2025-10-10 13:35:32.015524315 +0000 UTC m=+1045.913181078" Oct 10 13:35:32 crc kubenswrapper[4745]: I1010 13:35:32.374496 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 10 13:35:32 crc kubenswrapper[4745]: I1010 13:35:32.770758 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:35:32 crc kubenswrapper[4745]: I1010 13:35:32.956390 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b434a74-d01a-4af8-8d80-facf27b95128","Type":"ContainerStarted","Data":"57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e"} Oct 10 13:35:32 crc kubenswrapper[4745]: I1010 13:35:32.956785 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5b434a74-d01a-4af8-8d80-facf27b95128" containerName="nova-metadata-log" containerID="cri-o://33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead" gracePeriod=30 Oct 10 13:35:32 crc kubenswrapper[4745]: I1010 13:35:32.957225 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5b434a74-d01a-4af8-8d80-facf27b95128" containerName="nova-metadata-metadata" containerID="cri-o://57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e" gracePeriod=30 Oct 10 13:35:32 crc kubenswrapper[4745]: I1010 13:35:32.973850 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.036986198 podStartE2EDuration="5.973831454s" podCreationTimestamp="2025-10-10 13:35:27 +0000 UTC" firstStartedPulling="2025-10-10 13:35:28.285024314 +0000 UTC m=+1042.182681077" lastFinishedPulling="2025-10-10 13:35:31.22186957 +0000 UTC m=+1045.119526333" observedRunningTime="2025-10-10 13:35:32.973196038 +0000 UTC m=+1046.870852811" watchObservedRunningTime="2025-10-10 13:35:32.973831454 +0000 UTC m=+1046.871488217" Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.537353 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.643033 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b434a74-d01a-4af8-8d80-facf27b95128-config-data\") pod \"5b434a74-d01a-4af8-8d80-facf27b95128\" (UID: \"5b434a74-d01a-4af8-8d80-facf27b95128\") " Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.643266 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vxjs\" (UniqueName: \"kubernetes.io/projected/5b434a74-d01a-4af8-8d80-facf27b95128-kube-api-access-5vxjs\") pod \"5b434a74-d01a-4af8-8d80-facf27b95128\" (UID: \"5b434a74-d01a-4af8-8d80-facf27b95128\") " Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.643291 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b434a74-d01a-4af8-8d80-facf27b95128-combined-ca-bundle\") pod \"5b434a74-d01a-4af8-8d80-facf27b95128\" (UID: \"5b434a74-d01a-4af8-8d80-facf27b95128\") " Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.643354 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b434a74-d01a-4af8-8d80-facf27b95128-logs\") pod \"5b434a74-d01a-4af8-8d80-facf27b95128\" (UID: \"5b434a74-d01a-4af8-8d80-facf27b95128\") " Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.643884 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b434a74-d01a-4af8-8d80-facf27b95128-logs" (OuterVolumeSpecName: "logs") pod "5b434a74-d01a-4af8-8d80-facf27b95128" (UID: "5b434a74-d01a-4af8-8d80-facf27b95128"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.648952 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b434a74-d01a-4af8-8d80-facf27b95128-kube-api-access-5vxjs" (OuterVolumeSpecName: "kube-api-access-5vxjs") pod "5b434a74-d01a-4af8-8d80-facf27b95128" (UID: "5b434a74-d01a-4af8-8d80-facf27b95128"). InnerVolumeSpecName "kube-api-access-5vxjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.674604 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b434a74-d01a-4af8-8d80-facf27b95128-config-data" (OuterVolumeSpecName: "config-data") pod "5b434a74-d01a-4af8-8d80-facf27b95128" (UID: "5b434a74-d01a-4af8-8d80-facf27b95128"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.676416 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b434a74-d01a-4af8-8d80-facf27b95128-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b434a74-d01a-4af8-8d80-facf27b95128" (UID: "5b434a74-d01a-4af8-8d80-facf27b95128"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.745672 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b434a74-d01a-4af8-8d80-facf27b95128-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.745706 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b434a74-d01a-4af8-8d80-facf27b95128-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.745719 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vxjs\" (UniqueName: \"kubernetes.io/projected/5b434a74-d01a-4af8-8d80-facf27b95128-kube-api-access-5vxjs\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.745790 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b434a74-d01a-4af8-8d80-facf27b95128-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.981793 4745 generic.go:334] "Generic (PLEG): container finished" podID="5b434a74-d01a-4af8-8d80-facf27b95128" containerID="57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e" exitCode=0 Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.982846 4745 generic.go:334] "Generic (PLEG): container finished" podID="5b434a74-d01a-4af8-8d80-facf27b95128" containerID="33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead" exitCode=143 Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.981877 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.981899 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b434a74-d01a-4af8-8d80-facf27b95128","Type":"ContainerDied","Data":"57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e"} Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.985040 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b434a74-d01a-4af8-8d80-facf27b95128","Type":"ContainerDied","Data":"33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead"} Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.985071 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b434a74-d01a-4af8-8d80-facf27b95128","Type":"ContainerDied","Data":"68955bcdad9c5815104705d1cfbe83f6a47f577350468b0cb7ca5b66cd8eeba3"} Oct 10 13:35:33 crc kubenswrapper[4745]: I1010 13:35:33.985077 4745 scope.go:117] "RemoveContainer" containerID="57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.015965 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.031795 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.038431 4745 scope.go:117] "RemoveContainer" containerID="33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.066268 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:35:34 crc kubenswrapper[4745]: E1010 13:35:34.066782 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b434a74-d01a-4af8-8d80-facf27b95128" containerName="nova-metadata-metadata" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.066799 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b434a74-d01a-4af8-8d80-facf27b95128" containerName="nova-metadata-metadata" Oct 10 13:35:34 crc kubenswrapper[4745]: E1010 13:35:34.066825 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b434a74-d01a-4af8-8d80-facf27b95128" containerName="nova-metadata-log" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.066833 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b434a74-d01a-4af8-8d80-facf27b95128" containerName="nova-metadata-log" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.067059 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b434a74-d01a-4af8-8d80-facf27b95128" containerName="nova-metadata-log" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.067086 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b434a74-d01a-4af8-8d80-facf27b95128" containerName="nova-metadata-metadata" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.067137 4745 scope.go:117] "RemoveContainer" containerID="57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e" Oct 10 13:35:34 crc kubenswrapper[4745]: E1010 13:35:34.067902 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e\": container with ID starting with 57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e not found: ID does not exist" containerID="57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.067942 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e"} err="failed to get container status \"57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e\": rpc error: code = NotFound desc = could not find container \"57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e\": container with ID starting with 57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e not found: ID does not exist" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.067965 4745 scope.go:117] "RemoveContainer" containerID="33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.068174 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 13:35:34 crc kubenswrapper[4745]: E1010 13:35:34.068183 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead\": container with ID starting with 33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead not found: ID does not exist" containerID="33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.068229 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead"} err="failed to get container status \"33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead\": rpc error: code = NotFound desc = could not find container \"33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead\": container with ID starting with 33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead not found: ID does not exist" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.068242 4745 scope.go:117] "RemoveContainer" containerID="57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.068778 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e"} err="failed to get container status \"57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e\": rpc error: code = NotFound desc = could not find container \"57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e\": container with ID starting with 57efb550b635c8298e02755ac6f20e11a4fd4eed1272e787f29728864d62c73e not found: ID does not exist" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.068805 4745 scope.go:117] "RemoveContainer" containerID="33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.073506 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.073665 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.073909 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.075534 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead"} err="failed to get container status \"33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead\": rpc error: code = NotFound desc = could not find container \"33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead\": container with ID starting with 33e250a0eeea6a70c3a1307d64494001ad31d160e2dc7d446de67e716dc14ead not found: ID does not exist" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.170007 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " pod="openstack/nova-metadata-0" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.170191 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " pod="openstack/nova-metadata-0" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.170229 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-logs\") pod \"nova-metadata-0\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " pod="openstack/nova-metadata-0" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.170252 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-config-data\") pod \"nova-metadata-0\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " pod="openstack/nova-metadata-0" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.170301 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26g6s\" (UniqueName: \"kubernetes.io/projected/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-kube-api-access-26g6s\") pod \"nova-metadata-0\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " pod="openstack/nova-metadata-0" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.272010 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " pod="openstack/nova-metadata-0" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.272092 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-logs\") pod \"nova-metadata-0\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " pod="openstack/nova-metadata-0" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.272130 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-config-data\") pod \"nova-metadata-0\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " pod="openstack/nova-metadata-0" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.272208 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26g6s\" (UniqueName: \"kubernetes.io/projected/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-kube-api-access-26g6s\") pod \"nova-metadata-0\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " pod="openstack/nova-metadata-0" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.272324 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " pod="openstack/nova-metadata-0" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.273568 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-logs\") pod \"nova-metadata-0\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " pod="openstack/nova-metadata-0" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.277727 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " pod="openstack/nova-metadata-0" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.277973 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " pod="openstack/nova-metadata-0" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.290211 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-config-data\") pod \"nova-metadata-0\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " pod="openstack/nova-metadata-0" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.290714 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26g6s\" (UniqueName: \"kubernetes.io/projected/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-kube-api-access-26g6s\") pod \"nova-metadata-0\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " pod="openstack/nova-metadata-0" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.387078 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.760517 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b434a74-d01a-4af8-8d80-facf27b95128" path="/var/lib/kubelet/pods/5b434a74-d01a-4af8-8d80-facf27b95128/volumes" Oct 10 13:35:34 crc kubenswrapper[4745]: I1010 13:35:34.877926 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:35:34 crc kubenswrapper[4745]: W1010 13:35:34.881960 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8b0f383_ddb1_4d52_95b6_3d7a33c202ad.slice/crio-3bbd1b205f27c20dfe8a5b5fe09538daa21c83b71df391f8ea7ac75396aa1ef9 WatchSource:0}: Error finding container 3bbd1b205f27c20dfe8a5b5fe09538daa21c83b71df391f8ea7ac75396aa1ef9: Status 404 returned error can't find the container with id 3bbd1b205f27c20dfe8a5b5fe09538daa21c83b71df391f8ea7ac75396aa1ef9 Oct 10 13:35:35 crc kubenswrapper[4745]: I1010 13:35:35.009789 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad","Type":"ContainerStarted","Data":"3bbd1b205f27c20dfe8a5b5fe09538daa21c83b71df391f8ea7ac75396aa1ef9"} Oct 10 13:35:36 crc kubenswrapper[4745]: I1010 13:35:36.025088 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad","Type":"ContainerStarted","Data":"eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f"} Oct 10 13:35:36 crc kubenswrapper[4745]: I1010 13:35:36.025493 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad","Type":"ContainerStarted","Data":"b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724"} Oct 10 13:35:36 crc kubenswrapper[4745]: I1010 13:35:36.027285 4745 generic.go:334] "Generic (PLEG): container finished" podID="50933493-e7d7-4fc2-93bc-69173739fe58" containerID="41e66cac9137afc38339955332ce978e26364120893d9f3b693a00b8843750a2" exitCode=0 Oct 10 13:35:36 crc kubenswrapper[4745]: I1010 13:35:36.027395 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8t6r5" event={"ID":"50933493-e7d7-4fc2-93bc-69173739fe58","Type":"ContainerDied","Data":"41e66cac9137afc38339955332ce978e26364120893d9f3b693a00b8843750a2"} Oct 10 13:35:36 crc kubenswrapper[4745]: I1010 13:35:36.066262 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.066237248 podStartE2EDuration="2.066237248s" podCreationTimestamp="2025-10-10 13:35:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:35:36.064872163 +0000 UTC m=+1049.962528946" watchObservedRunningTime="2025-10-10 13:35:36.066237248 +0000 UTC m=+1049.963894021" Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.038586 4745 generic.go:334] "Generic (PLEG): container finished" podID="642aa23f-361f-4bc5-b463-15a104a8d7d7" containerID="7852ea056a15f2decd43c0c52a5fb22a6679a34eb1eb98afd32eba762c9edd28" exitCode=0 Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.038650 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6xlzw" event={"ID":"642aa23f-361f-4bc5-b463-15a104a8d7d7","Type":"ContainerDied","Data":"7852ea056a15f2decd43c0c52a5fb22a6679a34eb1eb98afd32eba762c9edd28"} Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.373699 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.405608 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.405849 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8t6r5" Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.474011 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.474632 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.529871 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcp7r\" (UniqueName: \"kubernetes.io/projected/50933493-e7d7-4fc2-93bc-69173739fe58-kube-api-access-kcp7r\") pod \"50933493-e7d7-4fc2-93bc-69173739fe58\" (UID: \"50933493-e7d7-4fc2-93bc-69173739fe58\") " Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.529991 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-scripts\") pod \"50933493-e7d7-4fc2-93bc-69173739fe58\" (UID: \"50933493-e7d7-4fc2-93bc-69173739fe58\") " Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.530122 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-combined-ca-bundle\") pod \"50933493-e7d7-4fc2-93bc-69173739fe58\" (UID: \"50933493-e7d7-4fc2-93bc-69173739fe58\") " Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.530255 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-config-data\") pod \"50933493-e7d7-4fc2-93bc-69173739fe58\" (UID: \"50933493-e7d7-4fc2-93bc-69173739fe58\") " Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.538171 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-scripts" (OuterVolumeSpecName: "scripts") pod "50933493-e7d7-4fc2-93bc-69173739fe58" (UID: "50933493-e7d7-4fc2-93bc-69173739fe58"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.538812 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50933493-e7d7-4fc2-93bc-69173739fe58-kube-api-access-kcp7r" (OuterVolumeSpecName: "kube-api-access-kcp7r") pod "50933493-e7d7-4fc2-93bc-69173739fe58" (UID: "50933493-e7d7-4fc2-93bc-69173739fe58"). InnerVolumeSpecName "kube-api-access-kcp7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.571390 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-config-data" (OuterVolumeSpecName: "config-data") pod "50933493-e7d7-4fc2-93bc-69173739fe58" (UID: "50933493-e7d7-4fc2-93bc-69173739fe58"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.581943 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50933493-e7d7-4fc2-93bc-69173739fe58" (UID: "50933493-e7d7-4fc2-93bc-69173739fe58"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.633708 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.633765 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcp7r\" (UniqueName: \"kubernetes.io/projected/50933493-e7d7-4fc2-93bc-69173739fe58-kube-api-access-kcp7r\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.633776 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.633785 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50933493-e7d7-4fc2-93bc-69173739fe58-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.794855 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.861549 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-rshbk"] Oct 10 13:35:37 crc kubenswrapper[4745]: I1010 13:35:37.861880 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-rshbk" podUID="16739af7-770b-4253-be87-ca62b9c21fb9" containerName="dnsmasq-dns" containerID="cri-o://638df36b11053557be882abd38eed1c627004848d3fe8dd43324de940253460a" gracePeriod=10 Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.049562 4745 generic.go:334] "Generic (PLEG): container finished" podID="16739af7-770b-4253-be87-ca62b9c21fb9" containerID="638df36b11053557be882abd38eed1c627004848d3fe8dd43324de940253460a" exitCode=0 Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.049627 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-rshbk" event={"ID":"16739af7-770b-4253-be87-ca62b9c21fb9","Type":"ContainerDied","Data":"638df36b11053557be882abd38eed1c627004848d3fe8dd43324de940253460a"} Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.052241 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8t6r5" event={"ID":"50933493-e7d7-4fc2-93bc-69173739fe58","Type":"ContainerDied","Data":"8ed91649e2c3326d702d4f20bf49662d2d6be31f8eff598ea62d90e8b2c4b687"} Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.052286 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ed91649e2c3326d702d4f20bf49662d2d6be31f8eff598ea62d90e8b2c4b687" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.052310 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8t6r5" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.097184 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.279685 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.307995 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.308192 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f8b0f383-ddb1-4d52-95b6-3d7a33c202ad" containerName="nova-metadata-log" containerID="cri-o://b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724" gracePeriod=30 Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.308356 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f8b0f383-ddb1-4d52-95b6-3d7a33c202ad" containerName="nova-metadata-metadata" containerID="cri-o://eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f" gracePeriod=30 Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.558991 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="24ee955e-b7b3-43c2-8dbe-7118951f037d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.559007 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="24ee955e-b7b3-43c2-8dbe-7118951f037d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.583000 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.657275 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6xlzw" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.658024 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.757557 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drtjp\" (UniqueName: \"kubernetes.io/projected/642aa23f-361f-4bc5-b463-15a104a8d7d7-kube-api-access-drtjp\") pod \"642aa23f-361f-4bc5-b463-15a104a8d7d7\" (UID: \"642aa23f-361f-4bc5-b463-15a104a8d7d7\") " Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.757594 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-ovsdbserver-nb\") pod \"16739af7-770b-4253-be87-ca62b9c21fb9\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.757652 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-combined-ca-bundle\") pod \"642aa23f-361f-4bc5-b463-15a104a8d7d7\" (UID: \"642aa23f-361f-4bc5-b463-15a104a8d7d7\") " Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.757678 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-config\") pod \"16739af7-770b-4253-be87-ca62b9c21fb9\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.757700 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-dns-swift-storage-0\") pod \"16739af7-770b-4253-be87-ca62b9c21fb9\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.757739 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-dns-svc\") pod \"16739af7-770b-4253-be87-ca62b9c21fb9\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.757771 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-ovsdbserver-sb\") pod \"16739af7-770b-4253-be87-ca62b9c21fb9\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.757858 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-scripts\") pod \"642aa23f-361f-4bc5-b463-15a104a8d7d7\" (UID: \"642aa23f-361f-4bc5-b463-15a104a8d7d7\") " Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.757918 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-config-data\") pod \"642aa23f-361f-4bc5-b463-15a104a8d7d7\" (UID: \"642aa23f-361f-4bc5-b463-15a104a8d7d7\") " Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.757994 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5h4c\" (UniqueName: \"kubernetes.io/projected/16739af7-770b-4253-be87-ca62b9c21fb9-kube-api-access-v5h4c\") pod \"16739af7-770b-4253-be87-ca62b9c21fb9\" (UID: \"16739af7-770b-4253-be87-ca62b9c21fb9\") " Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.767449 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16739af7-770b-4253-be87-ca62b9c21fb9-kube-api-access-v5h4c" (OuterVolumeSpecName: "kube-api-access-v5h4c") pod "16739af7-770b-4253-be87-ca62b9c21fb9" (UID: "16739af7-770b-4253-be87-ca62b9c21fb9"). InnerVolumeSpecName "kube-api-access-v5h4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.769904 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/642aa23f-361f-4bc5-b463-15a104a8d7d7-kube-api-access-drtjp" (OuterVolumeSpecName: "kube-api-access-drtjp") pod "642aa23f-361f-4bc5-b463-15a104a8d7d7" (UID: "642aa23f-361f-4bc5-b463-15a104a8d7d7"). InnerVolumeSpecName "kube-api-access-drtjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.778931 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.784581 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-scripts" (OuterVolumeSpecName: "scripts") pod "642aa23f-361f-4bc5-b463-15a104a8d7d7" (UID: "642aa23f-361f-4bc5-b463-15a104a8d7d7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.809213 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-config-data" (OuterVolumeSpecName: "config-data") pod "642aa23f-361f-4bc5-b463-15a104a8d7d7" (UID: "642aa23f-361f-4bc5-b463-15a104a8d7d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.821556 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "642aa23f-361f-4bc5-b463-15a104a8d7d7" (UID: "642aa23f-361f-4bc5-b463-15a104a8d7d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.833188 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "16739af7-770b-4253-be87-ca62b9c21fb9" (UID: "16739af7-770b-4253-be87-ca62b9c21fb9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.835388 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "16739af7-770b-4253-be87-ca62b9c21fb9" (UID: "16739af7-770b-4253-be87-ca62b9c21fb9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.836338 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "16739af7-770b-4253-be87-ca62b9c21fb9" (UID: "16739af7-770b-4253-be87-ca62b9c21fb9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.843827 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "16739af7-770b-4253-be87-ca62b9c21fb9" (UID: "16739af7-770b-4253-be87-ca62b9c21fb9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.844385 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-config" (OuterVolumeSpecName: "config") pod "16739af7-770b-4253-be87-ca62b9c21fb9" (UID: "16739af7-770b-4253-be87-ca62b9c21fb9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.860817 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-config-data\") pod \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.861063 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-logs\") pod \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.861269 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-nova-metadata-tls-certs\") pod \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.861382 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26g6s\" (UniqueName: \"kubernetes.io/projected/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-kube-api-access-26g6s\") pod \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.861507 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-combined-ca-bundle\") pod \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\" (UID: \"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad\") " Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.862203 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-logs" (OuterVolumeSpecName: "logs") pod "f8b0f383-ddb1-4d52-95b6-3d7a33c202ad" (UID: "f8b0f383-ddb1-4d52-95b6-3d7a33c202ad"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.863475 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5h4c\" (UniqueName: \"kubernetes.io/projected/16739af7-770b-4253-be87-ca62b9c21fb9-kube-api-access-v5h4c\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.864592 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drtjp\" (UniqueName: \"kubernetes.io/projected/642aa23f-361f-4bc5-b463-15a104a8d7d7-kube-api-access-drtjp\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.864629 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.864640 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.864654 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.864665 4745 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.864676 4745 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.864685 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16739af7-770b-4253-be87-ca62b9c21fb9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.864696 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.864707 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.864719 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/642aa23f-361f-4bc5-b463-15a104a8d7d7-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.871093 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-kube-api-access-26g6s" (OuterVolumeSpecName: "kube-api-access-26g6s") pod "f8b0f383-ddb1-4d52-95b6-3d7a33c202ad" (UID: "f8b0f383-ddb1-4d52-95b6-3d7a33c202ad"). InnerVolumeSpecName "kube-api-access-26g6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.890201 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8b0f383-ddb1-4d52-95b6-3d7a33c202ad" (UID: "f8b0f383-ddb1-4d52-95b6-3d7a33c202ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.890469 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-config-data" (OuterVolumeSpecName: "config-data") pod "f8b0f383-ddb1-4d52-95b6-3d7a33c202ad" (UID: "f8b0f383-ddb1-4d52-95b6-3d7a33c202ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.918981 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "f8b0f383-ddb1-4d52-95b6-3d7a33c202ad" (UID: "f8b0f383-ddb1-4d52-95b6-3d7a33c202ad"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.966333 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.966377 4745 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.966395 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26g6s\" (UniqueName: \"kubernetes.io/projected/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-kube-api-access-26g6s\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:38 crc kubenswrapper[4745]: I1010 13:35:38.966404 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.064266 4745 generic.go:334] "Generic (PLEG): container finished" podID="f8b0f383-ddb1-4d52-95b6-3d7a33c202ad" containerID="eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f" exitCode=0 Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.064303 4745 generic.go:334] "Generic (PLEG): container finished" podID="f8b0f383-ddb1-4d52-95b6-3d7a33c202ad" containerID="b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724" exitCode=143 Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.064334 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.064383 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad","Type":"ContainerDied","Data":"eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f"} Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.064411 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad","Type":"ContainerDied","Data":"b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724"} Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.064423 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8b0f383-ddb1-4d52-95b6-3d7a33c202ad","Type":"ContainerDied","Data":"3bbd1b205f27c20dfe8a5b5fe09538daa21c83b71df391f8ea7ac75396aa1ef9"} Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.064439 4745 scope.go:117] "RemoveContainer" containerID="eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.067450 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6xlzw" event={"ID":"642aa23f-361f-4bc5-b463-15a104a8d7d7","Type":"ContainerDied","Data":"e121e21ce5ead97ee421c2f9f8fc096225351410b7a910208ed835180c9a8fd4"} Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.067489 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e121e21ce5ead97ee421c2f9f8fc096225351410b7a910208ed835180c9a8fd4" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.067564 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6xlzw" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.072138 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-rshbk" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.074836 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-rshbk" event={"ID":"16739af7-770b-4253-be87-ca62b9c21fb9","Type":"ContainerDied","Data":"62ea94d9b8e4ace99183b3a7039e85d99945afc0610784d58569b0230c933e59"} Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.075283 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="24ee955e-b7b3-43c2-8dbe-7118951f037d" containerName="nova-api-log" containerID="cri-o://167f6a8d5aade40bd04faed225146f2986dc93ee34716a8fc5c88b1599d6d406" gracePeriod=30 Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.075435 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="24ee955e-b7b3-43c2-8dbe-7118951f037d" containerName="nova-api-api" containerID="cri-o://3e8c68e1a2099ddee002f06ea9351311f067ec5a8cfa088f36d3254930ad7d3b" gracePeriod=30 Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.129922 4745 scope.go:117] "RemoveContainer" containerID="b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.136529 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.153881 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.163391 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:35:39 crc kubenswrapper[4745]: E1010 13:35:39.164101 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8b0f383-ddb1-4d52-95b6-3d7a33c202ad" containerName="nova-metadata-metadata" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.164222 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8b0f383-ddb1-4d52-95b6-3d7a33c202ad" containerName="nova-metadata-metadata" Oct 10 13:35:39 crc kubenswrapper[4745]: E1010 13:35:39.164325 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642aa23f-361f-4bc5-b463-15a104a8d7d7" containerName="nova-cell1-conductor-db-sync" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.164400 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="642aa23f-361f-4bc5-b463-15a104a8d7d7" containerName="nova-cell1-conductor-db-sync" Oct 10 13:35:39 crc kubenswrapper[4745]: E1010 13:35:39.164489 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16739af7-770b-4253-be87-ca62b9c21fb9" containerName="init" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.164565 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="16739af7-770b-4253-be87-ca62b9c21fb9" containerName="init" Oct 10 13:35:39 crc kubenswrapper[4745]: E1010 13:35:39.164656 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50933493-e7d7-4fc2-93bc-69173739fe58" containerName="nova-manage" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.164750 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="50933493-e7d7-4fc2-93bc-69173739fe58" containerName="nova-manage" Oct 10 13:35:39 crc kubenswrapper[4745]: E1010 13:35:39.164835 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16739af7-770b-4253-be87-ca62b9c21fb9" containerName="dnsmasq-dns" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.164904 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="16739af7-770b-4253-be87-ca62b9c21fb9" containerName="dnsmasq-dns" Oct 10 13:35:39 crc kubenswrapper[4745]: E1010 13:35:39.164991 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8b0f383-ddb1-4d52-95b6-3d7a33c202ad" containerName="nova-metadata-log" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.165088 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8b0f383-ddb1-4d52-95b6-3d7a33c202ad" containerName="nova-metadata-log" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.165382 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8b0f383-ddb1-4d52-95b6-3d7a33c202ad" containerName="nova-metadata-metadata" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.165472 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="16739af7-770b-4253-be87-ca62b9c21fb9" containerName="dnsmasq-dns" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.165558 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="642aa23f-361f-4bc5-b463-15a104a8d7d7" containerName="nova-cell1-conductor-db-sync" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.165633 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="50933493-e7d7-4fc2-93bc-69173739fe58" containerName="nova-manage" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.165752 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8b0f383-ddb1-4d52-95b6-3d7a33c202ad" containerName="nova-metadata-log" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.167182 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.181891 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.183074 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.191103 4745 scope.go:117] "RemoveContainer" containerID="eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f" Oct 10 13:35:39 crc kubenswrapper[4745]: E1010 13:35:39.198457 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f\": container with ID starting with eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f not found: ID does not exist" containerID="eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.198523 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f"} err="failed to get container status \"eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f\": rpc error: code = NotFound desc = could not find container \"eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f\": container with ID starting with eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f not found: ID does not exist" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.198549 4745 scope.go:117] "RemoveContainer" containerID="b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724" Oct 10 13:35:39 crc kubenswrapper[4745]: E1010 13:35:39.200617 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724\": container with ID starting with b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724 not found: ID does not exist" containerID="b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.200644 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724"} err="failed to get container status \"b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724\": rpc error: code = NotFound desc = could not find container \"b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724\": container with ID starting with b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724 not found: ID does not exist" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.200660 4745 scope.go:117] "RemoveContainer" containerID="eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.200965 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f"} err="failed to get container status \"eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f\": rpc error: code = NotFound desc = could not find container \"eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f\": container with ID starting with eb05e01d4376d8586d2635eac56fa4b8de75f1bafc5407167823e603ce92d80f not found: ID does not exist" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.201114 4745 scope.go:117] "RemoveContainer" containerID="b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.201595 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724"} err="failed to get container status \"b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724\": rpc error: code = NotFound desc = could not find container \"b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724\": container with ID starting with b4666330a1341c652c1ba43afa905949a732190575cb314a6247c22a2ee6e724 not found: ID does not exist" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.201656 4745 scope.go:117] "RemoveContainer" containerID="638df36b11053557be882abd38eed1c627004848d3fe8dd43324de940253460a" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.204277 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.205512 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.213065 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.217989 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-rshbk"] Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.227386 4745 scope.go:117] "RemoveContainer" containerID="163f31cbd1e3a68a64787aa42c4d8bbff6ba6dc8f43f824d178a92cbea8d0032" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.227546 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-rshbk"] Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.243352 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.253777 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.273888 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.274138 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-logs\") pod \"nova-metadata-0\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.274243 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-config-data\") pod \"nova-metadata-0\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.274325 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87ac2f43-7e80-4310-8d3f-5750ae54cc06-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"87ac2f43-7e80-4310-8d3f-5750ae54cc06\") " pod="openstack/nova-cell1-conductor-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.274418 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxgc8\" (UniqueName: \"kubernetes.io/projected/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-kube-api-access-qxgc8\") pod \"nova-metadata-0\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.274931 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.275049 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klqsl\" (UniqueName: \"kubernetes.io/projected/87ac2f43-7e80-4310-8d3f-5750ae54cc06-kube-api-access-klqsl\") pod \"nova-cell1-conductor-0\" (UID: \"87ac2f43-7e80-4310-8d3f-5750ae54cc06\") " pod="openstack/nova-cell1-conductor-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.275117 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87ac2f43-7e80-4310-8d3f-5750ae54cc06-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"87ac2f43-7e80-4310-8d3f-5750ae54cc06\") " pod="openstack/nova-cell1-conductor-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.377086 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87ac2f43-7e80-4310-8d3f-5750ae54cc06-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"87ac2f43-7e80-4310-8d3f-5750ae54cc06\") " pod="openstack/nova-cell1-conductor-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.377668 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxgc8\" (UniqueName: \"kubernetes.io/projected/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-kube-api-access-qxgc8\") pod \"nova-metadata-0\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.377704 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.377796 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klqsl\" (UniqueName: \"kubernetes.io/projected/87ac2f43-7e80-4310-8d3f-5750ae54cc06-kube-api-access-klqsl\") pod \"nova-cell1-conductor-0\" (UID: \"87ac2f43-7e80-4310-8d3f-5750ae54cc06\") " pod="openstack/nova-cell1-conductor-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.377816 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87ac2f43-7e80-4310-8d3f-5750ae54cc06-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"87ac2f43-7e80-4310-8d3f-5750ae54cc06\") " pod="openstack/nova-cell1-conductor-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.377871 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.377970 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-logs\") pod \"nova-metadata-0\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.377995 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-config-data\") pod \"nova-metadata-0\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.378462 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-logs\") pod \"nova-metadata-0\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.382036 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87ac2f43-7e80-4310-8d3f-5750ae54cc06-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"87ac2f43-7e80-4310-8d3f-5750ae54cc06\") " pod="openstack/nova-cell1-conductor-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.383994 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.386534 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-config-data\") pod \"nova-metadata-0\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.387203 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87ac2f43-7e80-4310-8d3f-5750ae54cc06-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"87ac2f43-7e80-4310-8d3f-5750ae54cc06\") " pod="openstack/nova-cell1-conductor-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.387274 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.396150 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxgc8\" (UniqueName: \"kubernetes.io/projected/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-kube-api-access-qxgc8\") pod \"nova-metadata-0\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.399082 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klqsl\" (UniqueName: \"kubernetes.io/projected/87ac2f43-7e80-4310-8d3f-5750ae54cc06-kube-api-access-klqsl\") pod \"nova-cell1-conductor-0\" (UID: \"87ac2f43-7e80-4310-8d3f-5750ae54cc06\") " pod="openstack/nova-cell1-conductor-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.501482 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 13:35:39 crc kubenswrapper[4745]: I1010 13:35:39.523905 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 13:35:40 crc kubenswrapper[4745]: I1010 13:35:40.009994 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 13:35:40 crc kubenswrapper[4745]: I1010 13:35:40.023482 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:35:40 crc kubenswrapper[4745]: W1010 13:35:40.029563 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c8be38c_cfc0_4877_a6cb_297a8cc27c73.slice/crio-52f7a5169ea38de2643de6b330578618baced73141f976fa1c418d5981ffe250 WatchSource:0}: Error finding container 52f7a5169ea38de2643de6b330578618baced73141f976fa1c418d5981ffe250: Status 404 returned error can't find the container with id 52f7a5169ea38de2643de6b330578618baced73141f976fa1c418d5981ffe250 Oct 10 13:35:40 crc kubenswrapper[4745]: I1010 13:35:40.090074 4745 generic.go:334] "Generic (PLEG): container finished" podID="24ee955e-b7b3-43c2-8dbe-7118951f037d" containerID="167f6a8d5aade40bd04faed225146f2986dc93ee34716a8fc5c88b1599d6d406" exitCode=143 Oct 10 13:35:40 crc kubenswrapper[4745]: I1010 13:35:40.090150 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"24ee955e-b7b3-43c2-8dbe-7118951f037d","Type":"ContainerDied","Data":"167f6a8d5aade40bd04faed225146f2986dc93ee34716a8fc5c88b1599d6d406"} Oct 10 13:35:40 crc kubenswrapper[4745]: I1010 13:35:40.092102 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"87ac2f43-7e80-4310-8d3f-5750ae54cc06","Type":"ContainerStarted","Data":"8c8450434340377eaeebcc1762710bd97ee144d68a33f56a1da47e2ad90bd047"} Oct 10 13:35:40 crc kubenswrapper[4745]: I1010 13:35:40.093956 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0c8be38c-cfc0-4877-a6cb-297a8cc27c73","Type":"ContainerStarted","Data":"52f7a5169ea38de2643de6b330578618baced73141f976fa1c418d5981ffe250"} Oct 10 13:35:40 crc kubenswrapper[4745]: I1010 13:35:40.094071 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a65d55af-b9a5-4f6f-b62d-5683e387caa5" containerName="nova-scheduler-scheduler" containerID="cri-o://f673012d6f3cc5c2f4698094b96935ef912bb8c8392c8add4b3f559c0e551756" gracePeriod=30 Oct 10 13:35:40 crc kubenswrapper[4745]: I1010 13:35:40.754509 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16739af7-770b-4253-be87-ca62b9c21fb9" path="/var/lib/kubelet/pods/16739af7-770b-4253-be87-ca62b9c21fb9/volumes" Oct 10 13:35:40 crc kubenswrapper[4745]: I1010 13:35:40.755424 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8b0f383-ddb1-4d52-95b6-3d7a33c202ad" path="/var/lib/kubelet/pods/f8b0f383-ddb1-4d52-95b6-3d7a33c202ad/volumes" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.043067 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.103921 4745 generic.go:334] "Generic (PLEG): container finished" podID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerID="f31a93e54973a54b0849f6cdfcb1a8b5d9eaaeb94b47720c8b6a12c09a3f479e" exitCode=137 Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.103982 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fc8e852b-c861-4506-ab6f-9994a5cbe72d","Type":"ContainerDied","Data":"f31a93e54973a54b0849f6cdfcb1a8b5d9eaaeb94b47720c8b6a12c09a3f479e"} Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.104009 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fc8e852b-c861-4506-ab6f-9994a5cbe72d","Type":"ContainerDied","Data":"e39b428453da463ebeaae4f4185972c7b81891447f615593b03333fb0d634a99"} Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.104027 4745 scope.go:117] "RemoveContainer" containerID="f31a93e54973a54b0849f6cdfcb1a8b5d9eaaeb94b47720c8b6a12c09a3f479e" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.104157 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.107940 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0c8be38c-cfc0-4877-a6cb-297a8cc27c73","Type":"ContainerStarted","Data":"90bc1ee4e8020dc1894d4e1c0c175b564baa6d96f3f9d719e9e0e6b85a06e86c"} Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.108655 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0c8be38c-cfc0-4877-a6cb-297a8cc27c73","Type":"ContainerStarted","Data":"ad3fe94cc64220fec72e0955b52deac7a6ee30abed948ce06a3a7ef8cc04eea0"} Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.109696 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"87ac2f43-7e80-4310-8d3f-5750ae54cc06","Type":"ContainerStarted","Data":"f390cf735c1d7f379469070d44a26a19da999989e8efd896b9c1e34b46f1c1f1"} Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.109893 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.111640 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fc8e852b-c861-4506-ab6f-9994a5cbe72d-run-httpd\") pod \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.111748 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fc8e852b-c861-4506-ab6f-9994a5cbe72d-log-httpd\") pod \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.111908 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-sg-core-conf-yaml\") pod \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.111983 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-scripts\") pod \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.112022 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5p2q\" (UniqueName: \"kubernetes.io/projected/fc8e852b-c861-4506-ab6f-9994a5cbe72d-kube-api-access-c5p2q\") pod \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.112056 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-config-data\") pod \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.112116 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-combined-ca-bundle\") pod \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.112403 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc8e852b-c861-4506-ab6f-9994a5cbe72d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fc8e852b-c861-4506-ab6f-9994a5cbe72d" (UID: "fc8e852b-c861-4506-ab6f-9994a5cbe72d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.113444 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc8e852b-c861-4506-ab6f-9994a5cbe72d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fc8e852b-c861-4506-ab6f-9994a5cbe72d" (UID: "fc8e852b-c861-4506-ab6f-9994a5cbe72d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.114474 4745 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fc8e852b-c861-4506-ab6f-9994a5cbe72d-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.114504 4745 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fc8e852b-c861-4506-ab6f-9994a5cbe72d-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.117714 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-scripts" (OuterVolumeSpecName: "scripts") pod "fc8e852b-c861-4506-ab6f-9994a5cbe72d" (UID: "fc8e852b-c861-4506-ab6f-9994a5cbe72d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.117983 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc8e852b-c861-4506-ab6f-9994a5cbe72d-kube-api-access-c5p2q" (OuterVolumeSpecName: "kube-api-access-c5p2q") pod "fc8e852b-c861-4506-ab6f-9994a5cbe72d" (UID: "fc8e852b-c861-4506-ab6f-9994a5cbe72d"). InnerVolumeSpecName "kube-api-access-c5p2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.132549 4745 scope.go:117] "RemoveContainer" containerID="d342b70d70f95d528a84e1d535b35ea7acb11c89e7db5b4feca0931be5217a5c" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.134445 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.134427638 podStartE2EDuration="2.134427638s" podCreationTimestamp="2025-10-10 13:35:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:35:41.127194997 +0000 UTC m=+1055.024851770" watchObservedRunningTime="2025-10-10 13:35:41.134427638 +0000 UTC m=+1055.032084391" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.150098 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fc8e852b-c861-4506-ab6f-9994a5cbe72d" (UID: "fc8e852b-c861-4506-ab6f-9994a5cbe72d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.151895 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.151881234 podStartE2EDuration="2.151881234s" podCreationTimestamp="2025-10-10 13:35:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:35:41.147416713 +0000 UTC m=+1055.045073486" watchObservedRunningTime="2025-10-10 13:35:41.151881234 +0000 UTC m=+1055.049537997" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.200594 4745 scope.go:117] "RemoveContainer" containerID="d281eb8b85f7e3cedef128ce8c4d7bd51eab46d48fdbe85f4086e3d41c7bd806" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.203555 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc8e852b-c861-4506-ab6f-9994a5cbe72d" (UID: "fc8e852b-c861-4506-ab6f-9994a5cbe72d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.219079 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-config-data" (OuterVolumeSpecName: "config-data") pod "fc8e852b-c861-4506-ab6f-9994a5cbe72d" (UID: "fc8e852b-c861-4506-ab6f-9994a5cbe72d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.221565 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-config-data\") pod \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\" (UID: \"fc8e852b-c861-4506-ab6f-9994a5cbe72d\") " Oct 10 13:35:41 crc kubenswrapper[4745]: W1010 13:35:41.221717 4745 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/fc8e852b-c861-4506-ab6f-9994a5cbe72d/volumes/kubernetes.io~secret/config-data Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.221903 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-config-data" (OuterVolumeSpecName: "config-data") pod "fc8e852b-c861-4506-ab6f-9994a5cbe72d" (UID: "fc8e852b-c861-4506-ab6f-9994a5cbe72d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.223490 4745 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.223509 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.223519 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5p2q\" (UniqueName: \"kubernetes.io/projected/fc8e852b-c861-4506-ab6f-9994a5cbe72d-kube-api-access-c5p2q\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.223529 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.223538 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8e852b-c861-4506-ab6f-9994a5cbe72d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.231521 4745 scope.go:117] "RemoveContainer" containerID="43e933c97a8356b6a9f1f7a2b0719d9473dbb5ca5ced7154c6c3ad1e7cd4bb1d" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.249498 4745 scope.go:117] "RemoveContainer" containerID="f31a93e54973a54b0849f6cdfcb1a8b5d9eaaeb94b47720c8b6a12c09a3f479e" Oct 10 13:35:41 crc kubenswrapper[4745]: E1010 13:35:41.249925 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f31a93e54973a54b0849f6cdfcb1a8b5d9eaaeb94b47720c8b6a12c09a3f479e\": container with ID starting with f31a93e54973a54b0849f6cdfcb1a8b5d9eaaeb94b47720c8b6a12c09a3f479e not found: ID does not exist" containerID="f31a93e54973a54b0849f6cdfcb1a8b5d9eaaeb94b47720c8b6a12c09a3f479e" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.249977 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f31a93e54973a54b0849f6cdfcb1a8b5d9eaaeb94b47720c8b6a12c09a3f479e"} err="failed to get container status \"f31a93e54973a54b0849f6cdfcb1a8b5d9eaaeb94b47720c8b6a12c09a3f479e\": rpc error: code = NotFound desc = could not find container \"f31a93e54973a54b0849f6cdfcb1a8b5d9eaaeb94b47720c8b6a12c09a3f479e\": container with ID starting with f31a93e54973a54b0849f6cdfcb1a8b5d9eaaeb94b47720c8b6a12c09a3f479e not found: ID does not exist" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.250010 4745 scope.go:117] "RemoveContainer" containerID="d342b70d70f95d528a84e1d535b35ea7acb11c89e7db5b4feca0931be5217a5c" Oct 10 13:35:41 crc kubenswrapper[4745]: E1010 13:35:41.250339 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d342b70d70f95d528a84e1d535b35ea7acb11c89e7db5b4feca0931be5217a5c\": container with ID starting with d342b70d70f95d528a84e1d535b35ea7acb11c89e7db5b4feca0931be5217a5c not found: ID does not exist" containerID="d342b70d70f95d528a84e1d535b35ea7acb11c89e7db5b4feca0931be5217a5c" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.250365 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d342b70d70f95d528a84e1d535b35ea7acb11c89e7db5b4feca0931be5217a5c"} err="failed to get container status \"d342b70d70f95d528a84e1d535b35ea7acb11c89e7db5b4feca0931be5217a5c\": rpc error: code = NotFound desc = could not find container \"d342b70d70f95d528a84e1d535b35ea7acb11c89e7db5b4feca0931be5217a5c\": container with ID starting with d342b70d70f95d528a84e1d535b35ea7acb11c89e7db5b4feca0931be5217a5c not found: ID does not exist" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.250381 4745 scope.go:117] "RemoveContainer" containerID="d281eb8b85f7e3cedef128ce8c4d7bd51eab46d48fdbe85f4086e3d41c7bd806" Oct 10 13:35:41 crc kubenswrapper[4745]: E1010 13:35:41.250853 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d281eb8b85f7e3cedef128ce8c4d7bd51eab46d48fdbe85f4086e3d41c7bd806\": container with ID starting with d281eb8b85f7e3cedef128ce8c4d7bd51eab46d48fdbe85f4086e3d41c7bd806 not found: ID does not exist" containerID="d281eb8b85f7e3cedef128ce8c4d7bd51eab46d48fdbe85f4086e3d41c7bd806" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.250882 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d281eb8b85f7e3cedef128ce8c4d7bd51eab46d48fdbe85f4086e3d41c7bd806"} err="failed to get container status \"d281eb8b85f7e3cedef128ce8c4d7bd51eab46d48fdbe85f4086e3d41c7bd806\": rpc error: code = NotFound desc = could not find container \"d281eb8b85f7e3cedef128ce8c4d7bd51eab46d48fdbe85f4086e3d41c7bd806\": container with ID starting with d281eb8b85f7e3cedef128ce8c4d7bd51eab46d48fdbe85f4086e3d41c7bd806 not found: ID does not exist" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.250910 4745 scope.go:117] "RemoveContainer" containerID="43e933c97a8356b6a9f1f7a2b0719d9473dbb5ca5ced7154c6c3ad1e7cd4bb1d" Oct 10 13:35:41 crc kubenswrapper[4745]: E1010 13:35:41.251999 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43e933c97a8356b6a9f1f7a2b0719d9473dbb5ca5ced7154c6c3ad1e7cd4bb1d\": container with ID starting with 43e933c97a8356b6a9f1f7a2b0719d9473dbb5ca5ced7154c6c3ad1e7cd4bb1d not found: ID does not exist" containerID="43e933c97a8356b6a9f1f7a2b0719d9473dbb5ca5ced7154c6c3ad1e7cd4bb1d" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.252046 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43e933c97a8356b6a9f1f7a2b0719d9473dbb5ca5ced7154c6c3ad1e7cd4bb1d"} err="failed to get container status \"43e933c97a8356b6a9f1f7a2b0719d9473dbb5ca5ced7154c6c3ad1e7cd4bb1d\": rpc error: code = NotFound desc = could not find container \"43e933c97a8356b6a9f1f7a2b0719d9473dbb5ca5ced7154c6c3ad1e7cd4bb1d\": container with ID starting with 43e933c97a8356b6a9f1f7a2b0719d9473dbb5ca5ced7154c6c3ad1e7cd4bb1d not found: ID does not exist" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.461074 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.469198 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.495877 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:35:41 crc kubenswrapper[4745]: E1010 13:35:41.496581 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerName="ceilometer-notification-agent" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.496602 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerName="ceilometer-notification-agent" Oct 10 13:35:41 crc kubenswrapper[4745]: E1010 13:35:41.496629 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerName="sg-core" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.496637 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerName="sg-core" Oct 10 13:35:41 crc kubenswrapper[4745]: E1010 13:35:41.496653 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerName="ceilometer-central-agent" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.496661 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerName="ceilometer-central-agent" Oct 10 13:35:41 crc kubenswrapper[4745]: E1010 13:35:41.496712 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerName="proxy-httpd" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.496718 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerName="proxy-httpd" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.496902 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerName="proxy-httpd" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.496912 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerName="ceilometer-central-agent" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.496927 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerName="ceilometer-notification-agent" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.496941 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" containerName="sg-core" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.498679 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.500515 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.500514 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.515789 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.629937 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.630004 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-scripts\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.630111 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.630146 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n8wr\" (UniqueName: \"kubernetes.io/projected/a521b94a-74fa-4994-865d-69dc60b98ae2-kube-api-access-9n8wr\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.630171 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-config-data\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.630275 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a521b94a-74fa-4994-865d-69dc60b98ae2-log-httpd\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.630292 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a521b94a-74fa-4994-865d-69dc60b98ae2-run-httpd\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.732067 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.732156 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n8wr\" (UniqueName: \"kubernetes.io/projected/a521b94a-74fa-4994-865d-69dc60b98ae2-kube-api-access-9n8wr\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.732214 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-config-data\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.732264 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a521b94a-74fa-4994-865d-69dc60b98ae2-log-httpd\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.732296 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a521b94a-74fa-4994-865d-69dc60b98ae2-run-httpd\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.732430 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.732462 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-scripts\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.733063 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a521b94a-74fa-4994-865d-69dc60b98ae2-log-httpd\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.733232 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a521b94a-74fa-4994-865d-69dc60b98ae2-run-httpd\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.736016 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.737075 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.737580 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-config-data\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.749777 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-scripts\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.753171 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n8wr\" (UniqueName: \"kubernetes.io/projected/a521b94a-74fa-4994-865d-69dc60b98ae2-kube-api-access-9n8wr\") pod \"ceilometer-0\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " pod="openstack/ceilometer-0" Oct 10 13:35:41 crc kubenswrapper[4745]: I1010 13:35:41.827890 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:35:42 crc kubenswrapper[4745]: I1010 13:35:42.302408 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:35:42 crc kubenswrapper[4745]: W1010 13:35:42.306552 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda521b94a_74fa_4994_865d_69dc60b98ae2.slice/crio-77b6aa997d61b77dab352957d7dd427e508c8a663640088cd1071a1487fe73c0 WatchSource:0}: Error finding container 77b6aa997d61b77dab352957d7dd427e508c8a663640088cd1071a1487fe73c0: Status 404 returned error can't find the container with id 77b6aa997d61b77dab352957d7dd427e508c8a663640088cd1071a1487fe73c0 Oct 10 13:35:42 crc kubenswrapper[4745]: E1010 13:35:42.375782 4745 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f673012d6f3cc5c2f4698094b96935ef912bb8c8392c8add4b3f559c0e551756" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 13:35:42 crc kubenswrapper[4745]: E1010 13:35:42.377040 4745 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f673012d6f3cc5c2f4698094b96935ef912bb8c8392c8add4b3f559c0e551756" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 13:35:42 crc kubenswrapper[4745]: E1010 13:35:42.378867 4745 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f673012d6f3cc5c2f4698094b96935ef912bb8c8392c8add4b3f559c0e551756" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 13:35:42 crc kubenswrapper[4745]: E1010 13:35:42.378898 4745 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="a65d55af-b9a5-4f6f-b62d-5683e387caa5" containerName="nova-scheduler-scheduler" Oct 10 13:35:42 crc kubenswrapper[4745]: I1010 13:35:42.754383 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc8e852b-c861-4506-ab6f-9994a5cbe72d" path="/var/lib/kubelet/pods/fc8e852b-c861-4506-ab6f-9994a5cbe72d/volumes" Oct 10 13:35:43 crc kubenswrapper[4745]: I1010 13:35:43.130468 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a521b94a-74fa-4994-865d-69dc60b98ae2","Type":"ContainerStarted","Data":"77b6aa997d61b77dab352957d7dd427e508c8a663640088cd1071a1487fe73c0"} Oct 10 13:35:43 crc kubenswrapper[4745]: I1010 13:35:43.134049 4745 generic.go:334] "Generic (PLEG): container finished" podID="a65d55af-b9a5-4f6f-b62d-5683e387caa5" containerID="f673012d6f3cc5c2f4698094b96935ef912bb8c8392c8add4b3f559c0e551756" exitCode=0 Oct 10 13:35:43 crc kubenswrapper[4745]: I1010 13:35:43.134095 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a65d55af-b9a5-4f6f-b62d-5683e387caa5","Type":"ContainerDied","Data":"f673012d6f3cc5c2f4698094b96935ef912bb8c8392c8add4b3f559c0e551756"} Oct 10 13:35:43 crc kubenswrapper[4745]: I1010 13:35:43.360340 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 13:35:43 crc kubenswrapper[4745]: I1010 13:35:43.516523 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk8mr\" (UniqueName: \"kubernetes.io/projected/a65d55af-b9a5-4f6f-b62d-5683e387caa5-kube-api-access-tk8mr\") pod \"a65d55af-b9a5-4f6f-b62d-5683e387caa5\" (UID: \"a65d55af-b9a5-4f6f-b62d-5683e387caa5\") " Oct 10 13:35:43 crc kubenswrapper[4745]: I1010 13:35:43.516868 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a65d55af-b9a5-4f6f-b62d-5683e387caa5-config-data\") pod \"a65d55af-b9a5-4f6f-b62d-5683e387caa5\" (UID: \"a65d55af-b9a5-4f6f-b62d-5683e387caa5\") " Oct 10 13:35:43 crc kubenswrapper[4745]: I1010 13:35:43.517071 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a65d55af-b9a5-4f6f-b62d-5683e387caa5-combined-ca-bundle\") pod \"a65d55af-b9a5-4f6f-b62d-5683e387caa5\" (UID: \"a65d55af-b9a5-4f6f-b62d-5683e387caa5\") " Oct 10 13:35:43 crc kubenswrapper[4745]: I1010 13:35:43.523139 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a65d55af-b9a5-4f6f-b62d-5683e387caa5-kube-api-access-tk8mr" (OuterVolumeSpecName: "kube-api-access-tk8mr") pod "a65d55af-b9a5-4f6f-b62d-5683e387caa5" (UID: "a65d55af-b9a5-4f6f-b62d-5683e387caa5"). InnerVolumeSpecName "kube-api-access-tk8mr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:35:43 crc kubenswrapper[4745]: I1010 13:35:43.563445 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a65d55af-b9a5-4f6f-b62d-5683e387caa5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a65d55af-b9a5-4f6f-b62d-5683e387caa5" (UID: "a65d55af-b9a5-4f6f-b62d-5683e387caa5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:43 crc kubenswrapper[4745]: I1010 13:35:43.577480 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a65d55af-b9a5-4f6f-b62d-5683e387caa5-config-data" (OuterVolumeSpecName: "config-data") pod "a65d55af-b9a5-4f6f-b62d-5683e387caa5" (UID: "a65d55af-b9a5-4f6f-b62d-5683e387caa5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:43 crc kubenswrapper[4745]: I1010 13:35:43.620321 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk8mr\" (UniqueName: \"kubernetes.io/projected/a65d55af-b9a5-4f6f-b62d-5683e387caa5-kube-api-access-tk8mr\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:43 crc kubenswrapper[4745]: I1010 13:35:43.620364 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a65d55af-b9a5-4f6f-b62d-5683e387caa5-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:43 crc kubenswrapper[4745]: I1010 13:35:43.620378 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a65d55af-b9a5-4f6f-b62d-5683e387caa5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:43 crc kubenswrapper[4745]: I1010 13:35:43.854804 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.030798 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24ee955e-b7b3-43c2-8dbe-7118951f037d-logs\") pod \"24ee955e-b7b3-43c2-8dbe-7118951f037d\" (UID: \"24ee955e-b7b3-43c2-8dbe-7118951f037d\") " Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.031034 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frt9x\" (UniqueName: \"kubernetes.io/projected/24ee955e-b7b3-43c2-8dbe-7118951f037d-kube-api-access-frt9x\") pod \"24ee955e-b7b3-43c2-8dbe-7118951f037d\" (UID: \"24ee955e-b7b3-43c2-8dbe-7118951f037d\") " Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.031156 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ee955e-b7b3-43c2-8dbe-7118951f037d-config-data\") pod \"24ee955e-b7b3-43c2-8dbe-7118951f037d\" (UID: \"24ee955e-b7b3-43c2-8dbe-7118951f037d\") " Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.031257 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ee955e-b7b3-43c2-8dbe-7118951f037d-combined-ca-bundle\") pod \"24ee955e-b7b3-43c2-8dbe-7118951f037d\" (UID: \"24ee955e-b7b3-43c2-8dbe-7118951f037d\") " Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.034994 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24ee955e-b7b3-43c2-8dbe-7118951f037d-logs" (OuterVolumeSpecName: "logs") pod "24ee955e-b7b3-43c2-8dbe-7118951f037d" (UID: "24ee955e-b7b3-43c2-8dbe-7118951f037d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.037510 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24ee955e-b7b3-43c2-8dbe-7118951f037d-kube-api-access-frt9x" (OuterVolumeSpecName: "kube-api-access-frt9x") pod "24ee955e-b7b3-43c2-8dbe-7118951f037d" (UID: "24ee955e-b7b3-43c2-8dbe-7118951f037d"). InnerVolumeSpecName "kube-api-access-frt9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.058041 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24ee955e-b7b3-43c2-8dbe-7118951f037d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24ee955e-b7b3-43c2-8dbe-7118951f037d" (UID: "24ee955e-b7b3-43c2-8dbe-7118951f037d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.063693 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24ee955e-b7b3-43c2-8dbe-7118951f037d-config-data" (OuterVolumeSpecName: "config-data") pod "24ee955e-b7b3-43c2-8dbe-7118951f037d" (UID: "24ee955e-b7b3-43c2-8dbe-7118951f037d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.135028 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24ee955e-b7b3-43c2-8dbe-7118951f037d-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.135107 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frt9x\" (UniqueName: \"kubernetes.io/projected/24ee955e-b7b3-43c2-8dbe-7118951f037d-kube-api-access-frt9x\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.135124 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ee955e-b7b3-43c2-8dbe-7118951f037d-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.135141 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ee955e-b7b3-43c2-8dbe-7118951f037d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.149675 4745 generic.go:334] "Generic (PLEG): container finished" podID="24ee955e-b7b3-43c2-8dbe-7118951f037d" containerID="3e8c68e1a2099ddee002f06ea9351311f067ec5a8cfa088f36d3254930ad7d3b" exitCode=0 Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.149801 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"24ee955e-b7b3-43c2-8dbe-7118951f037d","Type":"ContainerDied","Data":"3e8c68e1a2099ddee002f06ea9351311f067ec5a8cfa088f36d3254930ad7d3b"} Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.149851 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"24ee955e-b7b3-43c2-8dbe-7118951f037d","Type":"ContainerDied","Data":"fe8feed6bed477636b2223879a7b8a40b77dba076a92233e691d67d7adbd6374"} Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.149877 4745 scope.go:117] "RemoveContainer" containerID="3e8c68e1a2099ddee002f06ea9351311f067ec5a8cfa088f36d3254930ad7d3b" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.150076 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.159717 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.159703 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a65d55af-b9a5-4f6f-b62d-5683e387caa5","Type":"ContainerDied","Data":"412d12e8fbc5c4c9ddd56c583e52695ad3a32af4cae8de8236336347dcc83dbc"} Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.169816 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a521b94a-74fa-4994-865d-69dc60b98ae2","Type":"ContainerStarted","Data":"df7ba72f56ab31ee5b5c05b3b86c49a19743512d1b47583733c548a45ff6fde3"} Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.169923 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a521b94a-74fa-4994-865d-69dc60b98ae2","Type":"ContainerStarted","Data":"5b03c9980967b61a0172950a7e7298ef1164a2539570758944df6298847fae37"} Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.205690 4745 scope.go:117] "RemoveContainer" containerID="167f6a8d5aade40bd04faed225146f2986dc93ee34716a8fc5c88b1599d6d406" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.208878 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.232345 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.267075 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.269484 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 10 13:35:44 crc kubenswrapper[4745]: E1010 13:35:44.269846 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24ee955e-b7b3-43c2-8dbe-7118951f037d" containerName="nova-api-log" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.269860 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="24ee955e-b7b3-43c2-8dbe-7118951f037d" containerName="nova-api-log" Oct 10 13:35:44 crc kubenswrapper[4745]: E1010 13:35:44.269883 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24ee955e-b7b3-43c2-8dbe-7118951f037d" containerName="nova-api-api" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.269888 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="24ee955e-b7b3-43c2-8dbe-7118951f037d" containerName="nova-api-api" Oct 10 13:35:44 crc kubenswrapper[4745]: E1010 13:35:44.269909 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a65d55af-b9a5-4f6f-b62d-5683e387caa5" containerName="nova-scheduler-scheduler" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.269916 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="a65d55af-b9a5-4f6f-b62d-5683e387caa5" containerName="nova-scheduler-scheduler" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.270094 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="24ee955e-b7b3-43c2-8dbe-7118951f037d" containerName="nova-api-log" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.270112 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="a65d55af-b9a5-4f6f-b62d-5683e387caa5" containerName="nova-scheduler-scheduler" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.270127 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="24ee955e-b7b3-43c2-8dbe-7118951f037d" containerName="nova-api-api" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.274913 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.278132 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.284565 4745 scope.go:117] "RemoveContainer" containerID="3e8c68e1a2099ddee002f06ea9351311f067ec5a8cfa088f36d3254930ad7d3b" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.290485 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 13:35:44 crc kubenswrapper[4745]: E1010 13:35:44.294288 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e8c68e1a2099ddee002f06ea9351311f067ec5a8cfa088f36d3254930ad7d3b\": container with ID starting with 3e8c68e1a2099ddee002f06ea9351311f067ec5a8cfa088f36d3254930ad7d3b not found: ID does not exist" containerID="3e8c68e1a2099ddee002f06ea9351311f067ec5a8cfa088f36d3254930ad7d3b" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.294527 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e8c68e1a2099ddee002f06ea9351311f067ec5a8cfa088f36d3254930ad7d3b"} err="failed to get container status \"3e8c68e1a2099ddee002f06ea9351311f067ec5a8cfa088f36d3254930ad7d3b\": rpc error: code = NotFound desc = could not find container \"3e8c68e1a2099ddee002f06ea9351311f067ec5a8cfa088f36d3254930ad7d3b\": container with ID starting with 3e8c68e1a2099ddee002f06ea9351311f067ec5a8cfa088f36d3254930ad7d3b not found: ID does not exist" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.294574 4745 scope.go:117] "RemoveContainer" containerID="167f6a8d5aade40bd04faed225146f2986dc93ee34716a8fc5c88b1599d6d406" Oct 10 13:35:44 crc kubenswrapper[4745]: E1010 13:35:44.296787 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"167f6a8d5aade40bd04faed225146f2986dc93ee34716a8fc5c88b1599d6d406\": container with ID starting with 167f6a8d5aade40bd04faed225146f2986dc93ee34716a8fc5c88b1599d6d406 not found: ID does not exist" containerID="167f6a8d5aade40bd04faed225146f2986dc93ee34716a8fc5c88b1599d6d406" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.296838 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"167f6a8d5aade40bd04faed225146f2986dc93ee34716a8fc5c88b1599d6d406"} err="failed to get container status \"167f6a8d5aade40bd04faed225146f2986dc93ee34716a8fc5c88b1599d6d406\": rpc error: code = NotFound desc = could not find container \"167f6a8d5aade40bd04faed225146f2986dc93ee34716a8fc5c88b1599d6d406\": container with ID starting with 167f6a8d5aade40bd04faed225146f2986dc93ee34716a8fc5c88b1599d6d406 not found: ID does not exist" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.296879 4745 scope.go:117] "RemoveContainer" containerID="f673012d6f3cc5c2f4698094b96935ef912bb8c8392c8add4b3f559c0e551756" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.316414 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.322364 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.324053 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.332352 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.352217 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.440664 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgjrz\" (UniqueName: \"kubernetes.io/projected/29c84d1a-c34f-4331-86c1-7c0f2184d98c-kube-api-access-zgjrz\") pod \"nova-scheduler-0\" (UID: \"29c84d1a-c34f-4331-86c1-7c0f2184d98c\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.440940 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-logs\") pod \"nova-api-0\" (UID: \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\") " pod="openstack/nova-api-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.441450 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-579ch\" (UniqueName: \"kubernetes.io/projected/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-kube-api-access-579ch\") pod \"nova-api-0\" (UID: \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\") " pod="openstack/nova-api-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.441509 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-config-data\") pod \"nova-api-0\" (UID: \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\") " pod="openstack/nova-api-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.441579 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\") " pod="openstack/nova-api-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.441600 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c84d1a-c34f-4331-86c1-7c0f2184d98c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"29c84d1a-c34f-4331-86c1-7c0f2184d98c\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.441817 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29c84d1a-c34f-4331-86c1-7c0f2184d98c-config-data\") pod \"nova-scheduler-0\" (UID: \"29c84d1a-c34f-4331-86c1-7c0f2184d98c\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.501912 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.502038 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.543032 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgjrz\" (UniqueName: \"kubernetes.io/projected/29c84d1a-c34f-4331-86c1-7c0f2184d98c-kube-api-access-zgjrz\") pod \"nova-scheduler-0\" (UID: \"29c84d1a-c34f-4331-86c1-7c0f2184d98c\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.543125 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-logs\") pod \"nova-api-0\" (UID: \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\") " pod="openstack/nova-api-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.543183 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-579ch\" (UniqueName: \"kubernetes.io/projected/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-kube-api-access-579ch\") pod \"nova-api-0\" (UID: \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\") " pod="openstack/nova-api-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.543201 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-config-data\") pod \"nova-api-0\" (UID: \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\") " pod="openstack/nova-api-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.543225 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\") " pod="openstack/nova-api-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.543240 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c84d1a-c34f-4331-86c1-7c0f2184d98c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"29c84d1a-c34f-4331-86c1-7c0f2184d98c\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.543260 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29c84d1a-c34f-4331-86c1-7c0f2184d98c-config-data\") pod \"nova-scheduler-0\" (UID: \"29c84d1a-c34f-4331-86c1-7c0f2184d98c\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.544804 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-logs\") pod \"nova-api-0\" (UID: \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\") " pod="openstack/nova-api-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.547229 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c84d1a-c34f-4331-86c1-7c0f2184d98c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"29c84d1a-c34f-4331-86c1-7c0f2184d98c\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.548413 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-config-data\") pod \"nova-api-0\" (UID: \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\") " pod="openstack/nova-api-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.556481 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29c84d1a-c34f-4331-86c1-7c0f2184d98c-config-data\") pod \"nova-scheduler-0\" (UID: \"29c84d1a-c34f-4331-86c1-7c0f2184d98c\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.561006 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-579ch\" (UniqueName: \"kubernetes.io/projected/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-kube-api-access-579ch\") pod \"nova-api-0\" (UID: \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\") " pod="openstack/nova-api-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.565050 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\") " pod="openstack/nova-api-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.567405 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgjrz\" (UniqueName: \"kubernetes.io/projected/29c84d1a-c34f-4331-86c1-7c0f2184d98c-kube-api-access-zgjrz\") pod \"nova-scheduler-0\" (UID: \"29c84d1a-c34f-4331-86c1-7c0f2184d98c\") " pod="openstack/nova-scheduler-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.602206 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.693509 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.763569 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24ee955e-b7b3-43c2-8dbe-7118951f037d" path="/var/lib/kubelet/pods/24ee955e-b7b3-43c2-8dbe-7118951f037d/volumes" Oct 10 13:35:44 crc kubenswrapper[4745]: I1010 13:35:44.764570 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a65d55af-b9a5-4f6f-b62d-5683e387caa5" path="/var/lib/kubelet/pods/a65d55af-b9a5-4f6f-b62d-5683e387caa5/volumes" Oct 10 13:35:45 crc kubenswrapper[4745]: I1010 13:35:45.058606 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 13:35:45 crc kubenswrapper[4745]: W1010 13:35:45.063948 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ce6355a_8ba4_4cee_aaae_ca798fe8a8ba.slice/crio-0ecda71fa12fded9b0585f9a41c6046343a703066f8ce96da132564d6c4bd80b WatchSource:0}: Error finding container 0ecda71fa12fded9b0585f9a41c6046343a703066f8ce96da132564d6c4bd80b: Status 404 returned error can't find the container with id 0ecda71fa12fded9b0585f9a41c6046343a703066f8ce96da132564d6c4bd80b Oct 10 13:35:45 crc kubenswrapper[4745]: I1010 13:35:45.161817 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 13:35:45 crc kubenswrapper[4745]: I1010 13:35:45.215313 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a521b94a-74fa-4994-865d-69dc60b98ae2","Type":"ContainerStarted","Data":"a7ff822044478b4838f4aa40cfb43671e2d7c830238e062b9ce1d85a368345c5"} Oct 10 13:35:45 crc kubenswrapper[4745]: I1010 13:35:45.219041 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba","Type":"ContainerStarted","Data":"0ecda71fa12fded9b0585f9a41c6046343a703066f8ce96da132564d6c4bd80b"} Oct 10 13:35:46 crc kubenswrapper[4745]: I1010 13:35:46.186568 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:35:46 crc kubenswrapper[4745]: I1010 13:35:46.187036 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:35:46 crc kubenswrapper[4745]: I1010 13:35:46.230018 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"29c84d1a-c34f-4331-86c1-7c0f2184d98c","Type":"ContainerStarted","Data":"5875dfe77b16322bc95114bd035d3465d421d7e7e7904dcb40e1431a18869ff1"} Oct 10 13:35:46 crc kubenswrapper[4745]: I1010 13:35:46.230060 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"29c84d1a-c34f-4331-86c1-7c0f2184d98c","Type":"ContainerStarted","Data":"56fde75b08cd25d7ae37f5af25a77b3ac8a497834a8afb06a8978b9d77750b52"} Oct 10 13:35:46 crc kubenswrapper[4745]: I1010 13:35:46.232160 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba","Type":"ContainerStarted","Data":"578ac938d2b45c6dbd96a5d95bee5672aee630897c0a88b58308d4124ce65fff"} Oct 10 13:35:46 crc kubenswrapper[4745]: I1010 13:35:46.232206 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba","Type":"ContainerStarted","Data":"17a2d68b64599e185794b7ef3cfeae4535394a21a468d0c5902c95ff8ec80d2e"} Oct 10 13:35:46 crc kubenswrapper[4745]: I1010 13:35:46.263546 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.2635236819999998 podStartE2EDuration="2.263523682s" podCreationTimestamp="2025-10-10 13:35:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:35:46.243148563 +0000 UTC m=+1060.140805326" watchObservedRunningTime="2025-10-10 13:35:46.263523682 +0000 UTC m=+1060.161180455" Oct 10 13:35:46 crc kubenswrapper[4745]: I1010 13:35:46.275127 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.2751092010000002 podStartE2EDuration="2.275109201s" podCreationTimestamp="2025-10-10 13:35:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:35:46.261297166 +0000 UTC m=+1060.158953939" watchObservedRunningTime="2025-10-10 13:35:46.275109201 +0000 UTC m=+1060.172765974" Oct 10 13:35:47 crc kubenswrapper[4745]: I1010 13:35:47.249479 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a521b94a-74fa-4994-865d-69dc60b98ae2","Type":"ContainerStarted","Data":"d9343510abce16e5d70e1d061bd7f0b4247a7bc5738ae266126acc896dd59404"} Oct 10 13:35:47 crc kubenswrapper[4745]: I1010 13:35:47.251401 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 13:35:47 crc kubenswrapper[4745]: I1010 13:35:47.292662 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.094666828 podStartE2EDuration="6.292635771s" podCreationTimestamp="2025-10-10 13:35:41 +0000 UTC" firstStartedPulling="2025-10-10 13:35:42.309803423 +0000 UTC m=+1056.207460186" lastFinishedPulling="2025-10-10 13:35:46.507772366 +0000 UTC m=+1060.405429129" observedRunningTime="2025-10-10 13:35:47.283036771 +0000 UTC m=+1061.180693534" watchObservedRunningTime="2025-10-10 13:35:47.292635771 +0000 UTC m=+1061.190292534" Oct 10 13:35:49 crc kubenswrapper[4745]: I1010 13:35:49.502686 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 13:35:49 crc kubenswrapper[4745]: I1010 13:35:49.503143 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 13:35:49 crc kubenswrapper[4745]: I1010 13:35:49.563451 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 10 13:35:49 crc kubenswrapper[4745]: I1010 13:35:49.693699 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 10 13:35:50 crc kubenswrapper[4745]: I1010 13:35:50.514851 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0c8be38c-cfc0-4877-a6cb-297a8cc27c73" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 13:35:50 crc kubenswrapper[4745]: I1010 13:35:50.514866 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0c8be38c-cfc0-4877-a6cb-297a8cc27c73" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 13:35:54 crc kubenswrapper[4745]: I1010 13:35:54.603073 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 13:35:54 crc kubenswrapper[4745]: I1010 13:35:54.603350 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 13:35:54 crc kubenswrapper[4745]: I1010 13:35:54.693698 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 10 13:35:54 crc kubenswrapper[4745]: I1010 13:35:54.742185 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 10 13:35:55 crc kubenswrapper[4745]: I1010 13:35:55.375942 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 10 13:35:55 crc kubenswrapper[4745]: I1010 13:35:55.686144 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 13:35:55 crc kubenswrapper[4745]: I1010 13:35:55.686167 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 13:35:59 crc kubenswrapper[4745]: I1010 13:35:59.508636 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 13:35:59 crc kubenswrapper[4745]: I1010 13:35:59.511309 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 13:35:59 crc kubenswrapper[4745]: I1010 13:35:59.521137 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 13:36:00 crc kubenswrapper[4745]: I1010 13:36:00.394850 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.398811 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.410650 4745 generic.go:334] "Generic (PLEG): container finished" podID="89b5e6b9-5dac-4856-83b3-fb61388231dc" containerID="3ec82c7f9d4232b42f193eea107a76ff7031e4f6c76dbcddbd6f83b0a7679d47" exitCode=137 Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.410696 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"89b5e6b9-5dac-4856-83b3-fb61388231dc","Type":"ContainerDied","Data":"3ec82c7f9d4232b42f193eea107a76ff7031e4f6c76dbcddbd6f83b0a7679d47"} Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.410762 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"89b5e6b9-5dac-4856-83b3-fb61388231dc","Type":"ContainerDied","Data":"a93cbae6081756558d97d56f5bc5eba4354a318671727c31cbaaca4575dfd3cc"} Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.410707 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.410780 4745 scope.go:117] "RemoveContainer" containerID="3ec82c7f9d4232b42f193eea107a76ff7031e4f6c76dbcddbd6f83b0a7679d47" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.442922 4745 scope.go:117] "RemoveContainer" containerID="3ec82c7f9d4232b42f193eea107a76ff7031e4f6c76dbcddbd6f83b0a7679d47" Oct 10 13:36:02 crc kubenswrapper[4745]: E1010 13:36:02.443384 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ec82c7f9d4232b42f193eea107a76ff7031e4f6c76dbcddbd6f83b0a7679d47\": container with ID starting with 3ec82c7f9d4232b42f193eea107a76ff7031e4f6c76dbcddbd6f83b0a7679d47 not found: ID does not exist" containerID="3ec82c7f9d4232b42f193eea107a76ff7031e4f6c76dbcddbd6f83b0a7679d47" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.443417 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ec82c7f9d4232b42f193eea107a76ff7031e4f6c76dbcddbd6f83b0a7679d47"} err="failed to get container status \"3ec82c7f9d4232b42f193eea107a76ff7031e4f6c76dbcddbd6f83b0a7679d47\": rpc error: code = NotFound desc = could not find container \"3ec82c7f9d4232b42f193eea107a76ff7031e4f6c76dbcddbd6f83b0a7679d47\": container with ID starting with 3ec82c7f9d4232b42f193eea107a76ff7031e4f6c76dbcddbd6f83b0a7679d47 not found: ID does not exist" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.509144 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b5e6b9-5dac-4856-83b3-fb61388231dc-combined-ca-bundle\") pod \"89b5e6b9-5dac-4856-83b3-fb61388231dc\" (UID: \"89b5e6b9-5dac-4856-83b3-fb61388231dc\") " Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.509487 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b5e6b9-5dac-4856-83b3-fb61388231dc-config-data\") pod \"89b5e6b9-5dac-4856-83b3-fb61388231dc\" (UID: \"89b5e6b9-5dac-4856-83b3-fb61388231dc\") " Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.509581 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77wps\" (UniqueName: \"kubernetes.io/projected/89b5e6b9-5dac-4856-83b3-fb61388231dc-kube-api-access-77wps\") pod \"89b5e6b9-5dac-4856-83b3-fb61388231dc\" (UID: \"89b5e6b9-5dac-4856-83b3-fb61388231dc\") " Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.515940 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89b5e6b9-5dac-4856-83b3-fb61388231dc-kube-api-access-77wps" (OuterVolumeSpecName: "kube-api-access-77wps") pod "89b5e6b9-5dac-4856-83b3-fb61388231dc" (UID: "89b5e6b9-5dac-4856-83b3-fb61388231dc"). InnerVolumeSpecName "kube-api-access-77wps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.535671 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b5e6b9-5dac-4856-83b3-fb61388231dc-config-data" (OuterVolumeSpecName: "config-data") pod "89b5e6b9-5dac-4856-83b3-fb61388231dc" (UID: "89b5e6b9-5dac-4856-83b3-fb61388231dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.558050 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b5e6b9-5dac-4856-83b3-fb61388231dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89b5e6b9-5dac-4856-83b3-fb61388231dc" (UID: "89b5e6b9-5dac-4856-83b3-fb61388231dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.611453 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b5e6b9-5dac-4856-83b3-fb61388231dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.611484 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b5e6b9-5dac-4856-83b3-fb61388231dc-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.611495 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77wps\" (UniqueName: \"kubernetes.io/projected/89b5e6b9-5dac-4856-83b3-fb61388231dc-kube-api-access-77wps\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.772174 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.788646 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.798835 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 13:36:02 crc kubenswrapper[4745]: E1010 13:36:02.799552 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b5e6b9-5dac-4856-83b3-fb61388231dc" containerName="nova-cell1-novncproxy-novncproxy" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.799595 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b5e6b9-5dac-4856-83b3-fb61388231dc" containerName="nova-cell1-novncproxy-novncproxy" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.799991 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="89b5e6b9-5dac-4856-83b3-fb61388231dc" containerName="nova-cell1-novncproxy-novncproxy" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.801070 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.804079 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.804504 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.804924 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.807525 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.917964 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ff9dc11-4059-440b-a21f-8dd6ef566d36-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ff9dc11-4059-440b-a21f-8dd6ef566d36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.918023 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ff9dc11-4059-440b-a21f-8dd6ef566d36-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ff9dc11-4059-440b-a21f-8dd6ef566d36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.918159 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ff9dc11-4059-440b-a21f-8dd6ef566d36-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ff9dc11-4059-440b-a21f-8dd6ef566d36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.918296 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8x7h\" (UniqueName: \"kubernetes.io/projected/6ff9dc11-4059-440b-a21f-8dd6ef566d36-kube-api-access-l8x7h\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ff9dc11-4059-440b-a21f-8dd6ef566d36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:02 crc kubenswrapper[4745]: I1010 13:36:02.918322 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ff9dc11-4059-440b-a21f-8dd6ef566d36-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ff9dc11-4059-440b-a21f-8dd6ef566d36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:03 crc kubenswrapper[4745]: I1010 13:36:03.019924 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ff9dc11-4059-440b-a21f-8dd6ef566d36-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ff9dc11-4059-440b-a21f-8dd6ef566d36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:03 crc kubenswrapper[4745]: I1010 13:36:03.020039 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8x7h\" (UniqueName: \"kubernetes.io/projected/6ff9dc11-4059-440b-a21f-8dd6ef566d36-kube-api-access-l8x7h\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ff9dc11-4059-440b-a21f-8dd6ef566d36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:03 crc kubenswrapper[4745]: I1010 13:36:03.020070 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ff9dc11-4059-440b-a21f-8dd6ef566d36-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ff9dc11-4059-440b-a21f-8dd6ef566d36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:03 crc kubenswrapper[4745]: I1010 13:36:03.020166 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ff9dc11-4059-440b-a21f-8dd6ef566d36-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ff9dc11-4059-440b-a21f-8dd6ef566d36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:03 crc kubenswrapper[4745]: I1010 13:36:03.020203 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ff9dc11-4059-440b-a21f-8dd6ef566d36-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ff9dc11-4059-440b-a21f-8dd6ef566d36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:03 crc kubenswrapper[4745]: I1010 13:36:03.023441 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ff9dc11-4059-440b-a21f-8dd6ef566d36-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ff9dc11-4059-440b-a21f-8dd6ef566d36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:03 crc kubenswrapper[4745]: I1010 13:36:03.023475 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ff9dc11-4059-440b-a21f-8dd6ef566d36-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ff9dc11-4059-440b-a21f-8dd6ef566d36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:03 crc kubenswrapper[4745]: I1010 13:36:03.024052 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ff9dc11-4059-440b-a21f-8dd6ef566d36-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ff9dc11-4059-440b-a21f-8dd6ef566d36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:03 crc kubenswrapper[4745]: I1010 13:36:03.024481 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ff9dc11-4059-440b-a21f-8dd6ef566d36-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ff9dc11-4059-440b-a21f-8dd6ef566d36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:03 crc kubenswrapper[4745]: I1010 13:36:03.039835 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8x7h\" (UniqueName: \"kubernetes.io/projected/6ff9dc11-4059-440b-a21f-8dd6ef566d36-kube-api-access-l8x7h\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ff9dc11-4059-440b-a21f-8dd6ef566d36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:03 crc kubenswrapper[4745]: I1010 13:36:03.135454 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:03 crc kubenswrapper[4745]: I1010 13:36:03.630644 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 13:36:04 crc kubenswrapper[4745]: I1010 13:36:04.439679 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6ff9dc11-4059-440b-a21f-8dd6ef566d36","Type":"ContainerStarted","Data":"952c126ce5f5b527649e396d727500caeec0a5c8d7d57468fa628d91496be0ed"} Oct 10 13:36:04 crc kubenswrapper[4745]: I1010 13:36:04.439946 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6ff9dc11-4059-440b-a21f-8dd6ef566d36","Type":"ContainerStarted","Data":"596696c2361ac242757d36f33850c5d98e907547c07d5866682539a7794801cb"} Oct 10 13:36:04 crc kubenswrapper[4745]: I1010 13:36:04.471708 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.471689065 podStartE2EDuration="2.471689065s" podCreationTimestamp="2025-10-10 13:36:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:36:04.464812033 +0000 UTC m=+1078.362468816" watchObservedRunningTime="2025-10-10 13:36:04.471689065 +0000 UTC m=+1078.369345838" Oct 10 13:36:04 crc kubenswrapper[4745]: I1010 13:36:04.610210 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 13:36:04 crc kubenswrapper[4745]: I1010 13:36:04.610797 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 13:36:04 crc kubenswrapper[4745]: I1010 13:36:04.610941 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 13:36:04 crc kubenswrapper[4745]: I1010 13:36:04.612763 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 13:36:04 crc kubenswrapper[4745]: I1010 13:36:04.766440 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89b5e6b9-5dac-4856-83b3-fb61388231dc" path="/var/lib/kubelet/pods/89b5e6b9-5dac-4856-83b3-fb61388231dc/volumes" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.458074 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.464131 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.675028 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-mlmn9"] Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.676602 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.695285 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2m4z\" (UniqueName: \"kubernetes.io/projected/cd189c07-04d2-4077-95bd-81e3504117cf-kube-api-access-x2m4z\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.695340 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-config\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.695375 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.695394 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.695412 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.695449 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.709691 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-mlmn9"] Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.797609 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2m4z\" (UniqueName: \"kubernetes.io/projected/cd189c07-04d2-4077-95bd-81e3504117cf-kube-api-access-x2m4z\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.797664 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-config\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.797701 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.797718 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.797751 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.797799 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.798676 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-config\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.798946 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.799115 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.799312 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.799840 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:05 crc kubenswrapper[4745]: I1010 13:36:05.816641 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2m4z\" (UniqueName: \"kubernetes.io/projected/cd189c07-04d2-4077-95bd-81e3504117cf-kube-api-access-x2m4z\") pod \"dnsmasq-dns-59cf4bdb65-mlmn9\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:06 crc kubenswrapper[4745]: I1010 13:36:06.019272 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:06 crc kubenswrapper[4745]: I1010 13:36:06.488683 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-mlmn9"] Oct 10 13:36:06 crc kubenswrapper[4745]: W1010 13:36:06.492225 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd189c07_04d2_4077_95bd_81e3504117cf.slice/crio-3693139174f559438f1673903e159ed79786e64b9118da704a0ced3d630b2ea3 WatchSource:0}: Error finding container 3693139174f559438f1673903e159ed79786e64b9118da704a0ced3d630b2ea3: Status 404 returned error can't find the container with id 3693139174f559438f1673903e159ed79786e64b9118da704a0ced3d630b2ea3 Oct 10 13:36:07 crc kubenswrapper[4745]: I1010 13:36:07.476591 4745 generic.go:334] "Generic (PLEG): container finished" podID="cd189c07-04d2-4077-95bd-81e3504117cf" containerID="3b6565cb907867419092b17abbf534adb76c6512d3d32baba23b0ceb87041078" exitCode=0 Oct 10 13:36:07 crc kubenswrapper[4745]: I1010 13:36:07.478668 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" event={"ID":"cd189c07-04d2-4077-95bd-81e3504117cf","Type":"ContainerDied","Data":"3b6565cb907867419092b17abbf534adb76c6512d3d32baba23b0ceb87041078"} Oct 10 13:36:07 crc kubenswrapper[4745]: I1010 13:36:07.478748 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" event={"ID":"cd189c07-04d2-4077-95bd-81e3504117cf","Type":"ContainerStarted","Data":"3693139174f559438f1673903e159ed79786e64b9118da704a0ced3d630b2ea3"} Oct 10 13:36:07 crc kubenswrapper[4745]: I1010 13:36:07.481682 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:36:07 crc kubenswrapper[4745]: I1010 13:36:07.482368 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerName="ceilometer-central-agent" containerID="cri-o://5b03c9980967b61a0172950a7e7298ef1164a2539570758944df6298847fae37" gracePeriod=30 Oct 10 13:36:07 crc kubenswrapper[4745]: I1010 13:36:07.482924 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerName="proxy-httpd" containerID="cri-o://d9343510abce16e5d70e1d061bd7f0b4247a7bc5738ae266126acc896dd59404" gracePeriod=30 Oct 10 13:36:07 crc kubenswrapper[4745]: I1010 13:36:07.482992 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerName="sg-core" containerID="cri-o://a7ff822044478b4838f4aa40cfb43671e2d7c830238e062b9ce1d85a368345c5" gracePeriod=30 Oct 10 13:36:07 crc kubenswrapper[4745]: I1010 13:36:07.483037 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerName="ceilometer-notification-agent" containerID="cri-o://df7ba72f56ab31ee5b5c05b3b86c49a19743512d1b47583733c548a45ff6fde3" gracePeriod=30 Oct 10 13:36:07 crc kubenswrapper[4745]: I1010 13:36:07.498011 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 10 13:36:08 crc kubenswrapper[4745]: I1010 13:36:08.135833 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:08 crc kubenswrapper[4745]: I1010 13:36:08.341467 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 13:36:08 crc kubenswrapper[4745]: I1010 13:36:08.485151 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" event={"ID":"cd189c07-04d2-4077-95bd-81e3504117cf","Type":"ContainerStarted","Data":"393276f07936731c5060a1e0cfe82ff3b7e2cd94fd1d6faf36b376ff5cb4f413"} Oct 10 13:36:08 crc kubenswrapper[4745]: I1010 13:36:08.485277 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:08 crc kubenswrapper[4745]: I1010 13:36:08.488255 4745 generic.go:334] "Generic (PLEG): container finished" podID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerID="d9343510abce16e5d70e1d061bd7f0b4247a7bc5738ae266126acc896dd59404" exitCode=0 Oct 10 13:36:08 crc kubenswrapper[4745]: I1010 13:36:08.488275 4745 generic.go:334] "Generic (PLEG): container finished" podID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerID="a7ff822044478b4838f4aa40cfb43671e2d7c830238e062b9ce1d85a368345c5" exitCode=2 Oct 10 13:36:08 crc kubenswrapper[4745]: I1010 13:36:08.488282 4745 generic.go:334] "Generic (PLEG): container finished" podID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerID="5b03c9980967b61a0172950a7e7298ef1164a2539570758944df6298847fae37" exitCode=0 Oct 10 13:36:08 crc kubenswrapper[4745]: I1010 13:36:08.488431 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba" containerName="nova-api-log" containerID="cri-o://17a2d68b64599e185794b7ef3cfeae4535394a21a468d0c5902c95ff8ec80d2e" gracePeriod=30 Oct 10 13:36:08 crc kubenswrapper[4745]: I1010 13:36:08.488612 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a521b94a-74fa-4994-865d-69dc60b98ae2","Type":"ContainerDied","Data":"d9343510abce16e5d70e1d061bd7f0b4247a7bc5738ae266126acc896dd59404"} Oct 10 13:36:08 crc kubenswrapper[4745]: I1010 13:36:08.488634 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a521b94a-74fa-4994-865d-69dc60b98ae2","Type":"ContainerDied","Data":"a7ff822044478b4838f4aa40cfb43671e2d7c830238e062b9ce1d85a368345c5"} Oct 10 13:36:08 crc kubenswrapper[4745]: I1010 13:36:08.488644 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a521b94a-74fa-4994-865d-69dc60b98ae2","Type":"ContainerDied","Data":"5b03c9980967b61a0172950a7e7298ef1164a2539570758944df6298847fae37"} Oct 10 13:36:08 crc kubenswrapper[4745]: I1010 13:36:08.488685 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba" containerName="nova-api-api" containerID="cri-o://578ac938d2b45c6dbd96a5d95bee5672aee630897c0a88b58308d4124ce65fff" gracePeriod=30 Oct 10 13:36:08 crc kubenswrapper[4745]: I1010 13:36:08.510382 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" podStartSLOduration=3.510365377 podStartE2EDuration="3.510365377s" podCreationTimestamp="2025-10-10 13:36:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:36:08.509936626 +0000 UTC m=+1082.407593389" watchObservedRunningTime="2025-10-10 13:36:08.510365377 +0000 UTC m=+1082.408022130" Oct 10 13:36:09 crc kubenswrapper[4745]: I1010 13:36:09.502786 4745 generic.go:334] "Generic (PLEG): container finished" podID="7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba" containerID="17a2d68b64599e185794b7ef3cfeae4535394a21a468d0c5902c95ff8ec80d2e" exitCode=143 Oct 10 13:36:09 crc kubenswrapper[4745]: I1010 13:36:09.502860 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba","Type":"ContainerDied","Data":"17a2d68b64599e185794b7ef3cfeae4535394a21a468d0c5902c95ff8ec80d2e"} Oct 10 13:36:09 crc kubenswrapper[4745]: I1010 13:36:09.993200 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.181353 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-config-data\") pod \"a521b94a-74fa-4994-865d-69dc60b98ae2\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.181581 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a521b94a-74fa-4994-865d-69dc60b98ae2-log-httpd\") pod \"a521b94a-74fa-4994-865d-69dc60b98ae2\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.181618 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-sg-core-conf-yaml\") pod \"a521b94a-74fa-4994-865d-69dc60b98ae2\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.181642 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n8wr\" (UniqueName: \"kubernetes.io/projected/a521b94a-74fa-4994-865d-69dc60b98ae2-kube-api-access-9n8wr\") pod \"a521b94a-74fa-4994-865d-69dc60b98ae2\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.181685 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-scripts\") pod \"a521b94a-74fa-4994-865d-69dc60b98ae2\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.181715 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a521b94a-74fa-4994-865d-69dc60b98ae2-run-httpd\") pod \"a521b94a-74fa-4994-865d-69dc60b98ae2\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.181771 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-combined-ca-bundle\") pod \"a521b94a-74fa-4994-865d-69dc60b98ae2\" (UID: \"a521b94a-74fa-4994-865d-69dc60b98ae2\") " Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.183919 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a521b94a-74fa-4994-865d-69dc60b98ae2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a521b94a-74fa-4994-865d-69dc60b98ae2" (UID: "a521b94a-74fa-4994-865d-69dc60b98ae2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.184572 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a521b94a-74fa-4994-865d-69dc60b98ae2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a521b94a-74fa-4994-865d-69dc60b98ae2" (UID: "a521b94a-74fa-4994-865d-69dc60b98ae2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.192605 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a521b94a-74fa-4994-865d-69dc60b98ae2-kube-api-access-9n8wr" (OuterVolumeSpecName: "kube-api-access-9n8wr") pod "a521b94a-74fa-4994-865d-69dc60b98ae2" (UID: "a521b94a-74fa-4994-865d-69dc60b98ae2"). InnerVolumeSpecName "kube-api-access-9n8wr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.196841 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-scripts" (OuterVolumeSpecName: "scripts") pod "a521b94a-74fa-4994-865d-69dc60b98ae2" (UID: "a521b94a-74fa-4994-865d-69dc60b98ae2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.238335 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a521b94a-74fa-4994-865d-69dc60b98ae2" (UID: "a521b94a-74fa-4994-865d-69dc60b98ae2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.286020 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.286049 4745 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a521b94a-74fa-4994-865d-69dc60b98ae2-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.286057 4745 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a521b94a-74fa-4994-865d-69dc60b98ae2-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.286065 4745 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.286075 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n8wr\" (UniqueName: \"kubernetes.io/projected/a521b94a-74fa-4994-865d-69dc60b98ae2-kube-api-access-9n8wr\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.308906 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a521b94a-74fa-4994-865d-69dc60b98ae2" (UID: "a521b94a-74fa-4994-865d-69dc60b98ae2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.334137 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-config-data" (OuterVolumeSpecName: "config-data") pod "a521b94a-74fa-4994-865d-69dc60b98ae2" (UID: "a521b94a-74fa-4994-865d-69dc60b98ae2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.389280 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.389320 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a521b94a-74fa-4994-865d-69dc60b98ae2-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.514966 4745 generic.go:334] "Generic (PLEG): container finished" podID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerID="df7ba72f56ab31ee5b5c05b3b86c49a19743512d1b47583733c548a45ff6fde3" exitCode=0 Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.515008 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a521b94a-74fa-4994-865d-69dc60b98ae2","Type":"ContainerDied","Data":"df7ba72f56ab31ee5b5c05b3b86c49a19743512d1b47583733c548a45ff6fde3"} Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.515021 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.515035 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a521b94a-74fa-4994-865d-69dc60b98ae2","Type":"ContainerDied","Data":"77b6aa997d61b77dab352957d7dd427e508c8a663640088cd1071a1487fe73c0"} Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.515052 4745 scope.go:117] "RemoveContainer" containerID="d9343510abce16e5d70e1d061bd7f0b4247a7bc5738ae266126acc896dd59404" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.547763 4745 scope.go:117] "RemoveContainer" containerID="a7ff822044478b4838f4aa40cfb43671e2d7c830238e062b9ce1d85a368345c5" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.551542 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.561545 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.569862 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:36:10 crc kubenswrapper[4745]: E1010 13:36:10.570265 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerName="ceilometer-central-agent" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.570282 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerName="ceilometer-central-agent" Oct 10 13:36:10 crc kubenswrapper[4745]: E1010 13:36:10.570304 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerName="sg-core" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.570310 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerName="sg-core" Oct 10 13:36:10 crc kubenswrapper[4745]: E1010 13:36:10.570328 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerName="proxy-httpd" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.570333 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerName="proxy-httpd" Oct 10 13:36:10 crc kubenswrapper[4745]: E1010 13:36:10.570348 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerName="ceilometer-notification-agent" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.570353 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerName="ceilometer-notification-agent" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.570508 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerName="sg-core" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.570527 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerName="ceilometer-notification-agent" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.570541 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerName="proxy-httpd" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.570539 4745 scope.go:117] "RemoveContainer" containerID="df7ba72f56ab31ee5b5c05b3b86c49a19743512d1b47583733c548a45ff6fde3" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.570549 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" containerName="ceilometer-central-agent" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.572801 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.576427 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.577349 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.598317 4745 scope.go:117] "RemoveContainer" containerID="5b03c9980967b61a0172950a7e7298ef1164a2539570758944df6298847fae37" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.628565 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.649559 4745 scope.go:117] "RemoveContainer" containerID="d9343510abce16e5d70e1d061bd7f0b4247a7bc5738ae266126acc896dd59404" Oct 10 13:36:10 crc kubenswrapper[4745]: E1010 13:36:10.649949 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9343510abce16e5d70e1d061bd7f0b4247a7bc5738ae266126acc896dd59404\": container with ID starting with d9343510abce16e5d70e1d061bd7f0b4247a7bc5738ae266126acc896dd59404 not found: ID does not exist" containerID="d9343510abce16e5d70e1d061bd7f0b4247a7bc5738ae266126acc896dd59404" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.649985 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9343510abce16e5d70e1d061bd7f0b4247a7bc5738ae266126acc896dd59404"} err="failed to get container status \"d9343510abce16e5d70e1d061bd7f0b4247a7bc5738ae266126acc896dd59404\": rpc error: code = NotFound desc = could not find container \"d9343510abce16e5d70e1d061bd7f0b4247a7bc5738ae266126acc896dd59404\": container with ID starting with d9343510abce16e5d70e1d061bd7f0b4247a7bc5738ae266126acc896dd59404 not found: ID does not exist" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.650010 4745 scope.go:117] "RemoveContainer" containerID="a7ff822044478b4838f4aa40cfb43671e2d7c830238e062b9ce1d85a368345c5" Oct 10 13:36:10 crc kubenswrapper[4745]: E1010 13:36:10.650296 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7ff822044478b4838f4aa40cfb43671e2d7c830238e062b9ce1d85a368345c5\": container with ID starting with a7ff822044478b4838f4aa40cfb43671e2d7c830238e062b9ce1d85a368345c5 not found: ID does not exist" containerID="a7ff822044478b4838f4aa40cfb43671e2d7c830238e062b9ce1d85a368345c5" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.650320 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7ff822044478b4838f4aa40cfb43671e2d7c830238e062b9ce1d85a368345c5"} err="failed to get container status \"a7ff822044478b4838f4aa40cfb43671e2d7c830238e062b9ce1d85a368345c5\": rpc error: code = NotFound desc = could not find container \"a7ff822044478b4838f4aa40cfb43671e2d7c830238e062b9ce1d85a368345c5\": container with ID starting with a7ff822044478b4838f4aa40cfb43671e2d7c830238e062b9ce1d85a368345c5 not found: ID does not exist" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.650337 4745 scope.go:117] "RemoveContainer" containerID="df7ba72f56ab31ee5b5c05b3b86c49a19743512d1b47583733c548a45ff6fde3" Oct 10 13:36:10 crc kubenswrapper[4745]: E1010 13:36:10.650586 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df7ba72f56ab31ee5b5c05b3b86c49a19743512d1b47583733c548a45ff6fde3\": container with ID starting with df7ba72f56ab31ee5b5c05b3b86c49a19743512d1b47583733c548a45ff6fde3 not found: ID does not exist" containerID="df7ba72f56ab31ee5b5c05b3b86c49a19743512d1b47583733c548a45ff6fde3" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.650611 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df7ba72f56ab31ee5b5c05b3b86c49a19743512d1b47583733c548a45ff6fde3"} err="failed to get container status \"df7ba72f56ab31ee5b5c05b3b86c49a19743512d1b47583733c548a45ff6fde3\": rpc error: code = NotFound desc = could not find container \"df7ba72f56ab31ee5b5c05b3b86c49a19743512d1b47583733c548a45ff6fde3\": container with ID starting with df7ba72f56ab31ee5b5c05b3b86c49a19743512d1b47583733c548a45ff6fde3 not found: ID does not exist" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.650642 4745 scope.go:117] "RemoveContainer" containerID="5b03c9980967b61a0172950a7e7298ef1164a2539570758944df6298847fae37" Oct 10 13:36:10 crc kubenswrapper[4745]: E1010 13:36:10.650879 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b03c9980967b61a0172950a7e7298ef1164a2539570758944df6298847fae37\": container with ID starting with 5b03c9980967b61a0172950a7e7298ef1164a2539570758944df6298847fae37 not found: ID does not exist" containerID="5b03c9980967b61a0172950a7e7298ef1164a2539570758944df6298847fae37" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.650907 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b03c9980967b61a0172950a7e7298ef1164a2539570758944df6298847fae37"} err="failed to get container status \"5b03c9980967b61a0172950a7e7298ef1164a2539570758944df6298847fae37\": rpc error: code = NotFound desc = could not find container \"5b03c9980967b61a0172950a7e7298ef1164a2539570758944df6298847fae37\": container with ID starting with 5b03c9980967b61a0172950a7e7298ef1164a2539570758944df6298847fae37 not found: ID does not exist" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.693580 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41ddef85-570e-4646-beed-88c7f0492ffc-log-httpd\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.693639 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.693657 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41ddef85-570e-4646-beed-88c7f0492ffc-run-httpd\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.693691 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-scripts\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.694400 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fswlz\" (UniqueName: \"kubernetes.io/projected/41ddef85-570e-4646-beed-88c7f0492ffc-kube-api-access-fswlz\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.694464 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-config-data\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.694482 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.773838 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a521b94a-74fa-4994-865d-69dc60b98ae2" path="/var/lib/kubelet/pods/a521b94a-74fa-4994-865d-69dc60b98ae2/volumes" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.796802 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-scripts\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.797240 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fswlz\" (UniqueName: \"kubernetes.io/projected/41ddef85-570e-4646-beed-88c7f0492ffc-kube-api-access-fswlz\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.797509 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-config-data\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.797559 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.797793 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41ddef85-570e-4646-beed-88c7f0492ffc-log-httpd\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.797901 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.797949 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41ddef85-570e-4646-beed-88c7f0492ffc-run-httpd\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.798488 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41ddef85-570e-4646-beed-88c7f0492ffc-log-httpd\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.798848 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41ddef85-570e-4646-beed-88c7f0492ffc-run-httpd\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.803036 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-config-data\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.803090 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-scripts\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.808172 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.809555 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.825364 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fswlz\" (UniqueName: \"kubernetes.io/projected/41ddef85-570e-4646-beed-88c7f0492ffc-kube-api-access-fswlz\") pod \"ceilometer-0\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " pod="openstack/ceilometer-0" Oct 10 13:36:10 crc kubenswrapper[4745]: I1010 13:36:10.890930 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:36:11 crc kubenswrapper[4745]: I1010 13:36:11.357929 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:36:11 crc kubenswrapper[4745]: W1010 13:36:11.364088 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41ddef85_570e_4646_beed_88c7f0492ffc.slice/crio-2110f4339ed875a72e5059cb55143b049e92787837626654fdc4c5595d5e833f WatchSource:0}: Error finding container 2110f4339ed875a72e5059cb55143b049e92787837626654fdc4c5595d5e833f: Status 404 returned error can't find the container with id 2110f4339ed875a72e5059cb55143b049e92787837626654fdc4c5595d5e833f Oct 10 13:36:11 crc kubenswrapper[4745]: I1010 13:36:11.551030 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41ddef85-570e-4646-beed-88c7f0492ffc","Type":"ContainerStarted","Data":"2110f4339ed875a72e5059cb55143b049e92787837626654fdc4c5595d5e833f"} Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.106429 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.230235 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-579ch\" (UniqueName: \"kubernetes.io/projected/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-kube-api-access-579ch\") pod \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\" (UID: \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\") " Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.230290 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-combined-ca-bundle\") pod \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\" (UID: \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\") " Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.230349 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-logs\") pod \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\" (UID: \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\") " Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.230506 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-config-data\") pod \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\" (UID: \"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba\") " Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.231787 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-logs" (OuterVolumeSpecName: "logs") pod "7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba" (UID: "7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.238975 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-kube-api-access-579ch" (OuterVolumeSpecName: "kube-api-access-579ch") pod "7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba" (UID: "7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba"). InnerVolumeSpecName "kube-api-access-579ch". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.266661 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba" (UID: "7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.269271 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-config-data" (OuterVolumeSpecName: "config-data") pod "7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba" (UID: "7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.332534 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-579ch\" (UniqueName: \"kubernetes.io/projected/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-kube-api-access-579ch\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.332570 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.332580 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.332591 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.560384 4745 generic.go:334] "Generic (PLEG): container finished" podID="7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba" containerID="578ac938d2b45c6dbd96a5d95bee5672aee630897c0a88b58308d4124ce65fff" exitCode=0 Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.560444 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba","Type":"ContainerDied","Data":"578ac938d2b45c6dbd96a5d95bee5672aee630897c0a88b58308d4124ce65fff"} Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.560519 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba","Type":"ContainerDied","Data":"0ecda71fa12fded9b0585f9a41c6046343a703066f8ce96da132564d6c4bd80b"} Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.560555 4745 scope.go:117] "RemoveContainer" containerID="578ac938d2b45c6dbd96a5d95bee5672aee630897c0a88b58308d4124ce65fff" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.561405 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.561951 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41ddef85-570e-4646-beed-88c7f0492ffc","Type":"ContainerStarted","Data":"57eafa639f92ba605f2bdef31e303ca544d927866b6c441756fe1ab099939887"} Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.595915 4745 scope.go:117] "RemoveContainer" containerID="17a2d68b64599e185794b7ef3cfeae4535394a21a468d0c5902c95ff8ec80d2e" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.610271 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.615838 4745 scope.go:117] "RemoveContainer" containerID="578ac938d2b45c6dbd96a5d95bee5672aee630897c0a88b58308d4124ce65fff" Oct 10 13:36:12 crc kubenswrapper[4745]: E1010 13:36:12.616388 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"578ac938d2b45c6dbd96a5d95bee5672aee630897c0a88b58308d4124ce65fff\": container with ID starting with 578ac938d2b45c6dbd96a5d95bee5672aee630897c0a88b58308d4124ce65fff not found: ID does not exist" containerID="578ac938d2b45c6dbd96a5d95bee5672aee630897c0a88b58308d4124ce65fff" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.616433 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"578ac938d2b45c6dbd96a5d95bee5672aee630897c0a88b58308d4124ce65fff"} err="failed to get container status \"578ac938d2b45c6dbd96a5d95bee5672aee630897c0a88b58308d4124ce65fff\": rpc error: code = NotFound desc = could not find container \"578ac938d2b45c6dbd96a5d95bee5672aee630897c0a88b58308d4124ce65fff\": container with ID starting with 578ac938d2b45c6dbd96a5d95bee5672aee630897c0a88b58308d4124ce65fff not found: ID does not exist" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.616467 4745 scope.go:117] "RemoveContainer" containerID="17a2d68b64599e185794b7ef3cfeae4535394a21a468d0c5902c95ff8ec80d2e" Oct 10 13:36:12 crc kubenswrapper[4745]: E1010 13:36:12.616851 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17a2d68b64599e185794b7ef3cfeae4535394a21a468d0c5902c95ff8ec80d2e\": container with ID starting with 17a2d68b64599e185794b7ef3cfeae4535394a21a468d0c5902c95ff8ec80d2e not found: ID does not exist" containerID="17a2d68b64599e185794b7ef3cfeae4535394a21a468d0c5902c95ff8ec80d2e" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.616876 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17a2d68b64599e185794b7ef3cfeae4535394a21a468d0c5902c95ff8ec80d2e"} err="failed to get container status \"17a2d68b64599e185794b7ef3cfeae4535394a21a468d0c5902c95ff8ec80d2e\": rpc error: code = NotFound desc = could not find container \"17a2d68b64599e185794b7ef3cfeae4535394a21a468d0c5902c95ff8ec80d2e\": container with ID starting with 17a2d68b64599e185794b7ef3cfeae4535394a21a468d0c5902c95ff8ec80d2e not found: ID does not exist" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.627707 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.639148 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 10 13:36:12 crc kubenswrapper[4745]: E1010 13:36:12.639954 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba" containerName="nova-api-log" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.639985 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba" containerName="nova-api-log" Oct 10 13:36:12 crc kubenswrapper[4745]: E1010 13:36:12.639999 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba" containerName="nova-api-api" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.640005 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba" containerName="nova-api-api" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.640199 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba" containerName="nova-api-api" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.640234 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba" containerName="nova-api-log" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.643703 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.645760 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.646165 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.646227 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.659502 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.738347 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.738402 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/085f48ec-754f-4822-a379-b6fb7e4c44b3-logs\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.738432 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rspmz\" (UniqueName: \"kubernetes.io/projected/085f48ec-754f-4822-a379-b6fb7e4c44b3-kube-api-access-rspmz\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.738585 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-config-data\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.738678 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.738788 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-public-tls-certs\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.761039 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba" path="/var/lib/kubelet/pods/7ce6355a-8ba4-4cee-aaae-ca798fe8a8ba/volumes" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.842518 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.842605 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/085f48ec-754f-4822-a379-b6fb7e4c44b3-logs\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.842924 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rspmz\" (UniqueName: \"kubernetes.io/projected/085f48ec-754f-4822-a379-b6fb7e4c44b3-kube-api-access-rspmz\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.843070 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-config-data\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.843144 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.843195 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-public-tls-certs\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.843898 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/085f48ec-754f-4822-a379-b6fb7e4c44b3-logs\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.849303 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-public-tls-certs\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.849828 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-config-data\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.851239 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.852644 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:12 crc kubenswrapper[4745]: I1010 13:36:12.861009 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rspmz\" (UniqueName: \"kubernetes.io/projected/085f48ec-754f-4822-a379-b6fb7e4c44b3-kube-api-access-rspmz\") pod \"nova-api-0\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " pod="openstack/nova-api-0" Oct 10 13:36:13 crc kubenswrapper[4745]: I1010 13:36:13.136419 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:13 crc kubenswrapper[4745]: I1010 13:36:13.144391 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 13:36:13 crc kubenswrapper[4745]: I1010 13:36:13.167887 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:13 crc kubenswrapper[4745]: I1010 13:36:13.576752 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 13:36:13 crc kubenswrapper[4745]: I1010 13:36:13.581032 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41ddef85-570e-4646-beed-88c7f0492ffc","Type":"ContainerStarted","Data":"03d239f2c573a14d66f379afaf5947b1ef2c2196749ecf9c506ce139bd438b94"} Oct 10 13:36:13 crc kubenswrapper[4745]: I1010 13:36:13.609225 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 10 13:36:13 crc kubenswrapper[4745]: I1010 13:36:13.842655 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-ctcrf"] Oct 10 13:36:13 crc kubenswrapper[4745]: I1010 13:36:13.852183 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ctcrf" Oct 10 13:36:13 crc kubenswrapper[4745]: I1010 13:36:13.855551 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 10 13:36:13 crc kubenswrapper[4745]: I1010 13:36:13.858650 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 10 13:36:13 crc kubenswrapper[4745]: I1010 13:36:13.869609 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-ctcrf"] Oct 10 13:36:13 crc kubenswrapper[4745]: I1010 13:36:13.897627 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-config-data\") pod \"nova-cell1-cell-mapping-ctcrf\" (UID: \"16ee733c-3513-40bc-97ec-515a9f735a56\") " pod="openstack/nova-cell1-cell-mapping-ctcrf" Oct 10 13:36:13 crc kubenswrapper[4745]: I1010 13:36:13.897680 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ctcrf\" (UID: \"16ee733c-3513-40bc-97ec-515a9f735a56\") " pod="openstack/nova-cell1-cell-mapping-ctcrf" Oct 10 13:36:13 crc kubenswrapper[4745]: I1010 13:36:13.897962 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgcf4\" (UniqueName: \"kubernetes.io/projected/16ee733c-3513-40bc-97ec-515a9f735a56-kube-api-access-pgcf4\") pod \"nova-cell1-cell-mapping-ctcrf\" (UID: \"16ee733c-3513-40bc-97ec-515a9f735a56\") " pod="openstack/nova-cell1-cell-mapping-ctcrf" Oct 10 13:36:13 crc kubenswrapper[4745]: I1010 13:36:13.898098 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-scripts\") pod \"nova-cell1-cell-mapping-ctcrf\" (UID: \"16ee733c-3513-40bc-97ec-515a9f735a56\") " pod="openstack/nova-cell1-cell-mapping-ctcrf" Oct 10 13:36:13 crc kubenswrapper[4745]: I1010 13:36:13.999700 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-scripts\") pod \"nova-cell1-cell-mapping-ctcrf\" (UID: \"16ee733c-3513-40bc-97ec-515a9f735a56\") " pod="openstack/nova-cell1-cell-mapping-ctcrf" Oct 10 13:36:14 crc kubenswrapper[4745]: I1010 13:36:13.999867 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-config-data\") pod \"nova-cell1-cell-mapping-ctcrf\" (UID: \"16ee733c-3513-40bc-97ec-515a9f735a56\") " pod="openstack/nova-cell1-cell-mapping-ctcrf" Oct 10 13:36:14 crc kubenswrapper[4745]: I1010 13:36:13.999885 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ctcrf\" (UID: \"16ee733c-3513-40bc-97ec-515a9f735a56\") " pod="openstack/nova-cell1-cell-mapping-ctcrf" Oct 10 13:36:14 crc kubenswrapper[4745]: I1010 13:36:14.000570 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgcf4\" (UniqueName: \"kubernetes.io/projected/16ee733c-3513-40bc-97ec-515a9f735a56-kube-api-access-pgcf4\") pod \"nova-cell1-cell-mapping-ctcrf\" (UID: \"16ee733c-3513-40bc-97ec-515a9f735a56\") " pod="openstack/nova-cell1-cell-mapping-ctcrf" Oct 10 13:36:14 crc kubenswrapper[4745]: I1010 13:36:14.003400 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-config-data\") pod \"nova-cell1-cell-mapping-ctcrf\" (UID: \"16ee733c-3513-40bc-97ec-515a9f735a56\") " pod="openstack/nova-cell1-cell-mapping-ctcrf" Oct 10 13:36:14 crc kubenswrapper[4745]: I1010 13:36:14.004383 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-scripts\") pod \"nova-cell1-cell-mapping-ctcrf\" (UID: \"16ee733c-3513-40bc-97ec-515a9f735a56\") " pod="openstack/nova-cell1-cell-mapping-ctcrf" Oct 10 13:36:14 crc kubenswrapper[4745]: I1010 13:36:14.007455 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ctcrf\" (UID: \"16ee733c-3513-40bc-97ec-515a9f735a56\") " pod="openstack/nova-cell1-cell-mapping-ctcrf" Oct 10 13:36:14 crc kubenswrapper[4745]: I1010 13:36:14.024570 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgcf4\" (UniqueName: \"kubernetes.io/projected/16ee733c-3513-40bc-97ec-515a9f735a56-kube-api-access-pgcf4\") pod \"nova-cell1-cell-mapping-ctcrf\" (UID: \"16ee733c-3513-40bc-97ec-515a9f735a56\") " pod="openstack/nova-cell1-cell-mapping-ctcrf" Oct 10 13:36:14 crc kubenswrapper[4745]: I1010 13:36:14.047759 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ctcrf" Oct 10 13:36:14 crc kubenswrapper[4745]: I1010 13:36:14.492962 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-ctcrf"] Oct 10 13:36:14 crc kubenswrapper[4745]: W1010 13:36:14.497626 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16ee733c_3513_40bc_97ec_515a9f735a56.slice/crio-36870fbeb7dc4c49dff284ca7cf53df3e52b38ea09feaac6149a58622e83698e WatchSource:0}: Error finding container 36870fbeb7dc4c49dff284ca7cf53df3e52b38ea09feaac6149a58622e83698e: Status 404 returned error can't find the container with id 36870fbeb7dc4c49dff284ca7cf53df3e52b38ea09feaac6149a58622e83698e Oct 10 13:36:14 crc kubenswrapper[4745]: I1010 13:36:14.606165 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41ddef85-570e-4646-beed-88c7f0492ffc","Type":"ContainerStarted","Data":"38d579f086d9c930d1b9219cb7e9cdbaa805b2fc2db70b85854c7b6f5ff62d14"} Oct 10 13:36:14 crc kubenswrapper[4745]: I1010 13:36:14.608569 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"085f48ec-754f-4822-a379-b6fb7e4c44b3","Type":"ContainerStarted","Data":"ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96"} Oct 10 13:36:14 crc kubenswrapper[4745]: I1010 13:36:14.608610 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"085f48ec-754f-4822-a379-b6fb7e4c44b3","Type":"ContainerStarted","Data":"02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042"} Oct 10 13:36:14 crc kubenswrapper[4745]: I1010 13:36:14.608623 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"085f48ec-754f-4822-a379-b6fb7e4c44b3","Type":"ContainerStarted","Data":"ffab2c4096dd0039e5b1a0164c981aa75dd24e98b88a632176cd77e53329c1e1"} Oct 10 13:36:14 crc kubenswrapper[4745]: I1010 13:36:14.610607 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ctcrf" event={"ID":"16ee733c-3513-40bc-97ec-515a9f735a56","Type":"ContainerStarted","Data":"36870fbeb7dc4c49dff284ca7cf53df3e52b38ea09feaac6149a58622e83698e"} Oct 10 13:36:14 crc kubenswrapper[4745]: I1010 13:36:14.633961 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.633944613 podStartE2EDuration="2.633944613s" podCreationTimestamp="2025-10-10 13:36:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:36:14.62983022 +0000 UTC m=+1088.527487003" watchObservedRunningTime="2025-10-10 13:36:14.633944613 +0000 UTC m=+1088.531601376" Oct 10 13:36:15 crc kubenswrapper[4745]: I1010 13:36:15.622164 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ctcrf" event={"ID":"16ee733c-3513-40bc-97ec-515a9f735a56","Type":"ContainerStarted","Data":"2e6f2227fb0e20b163ece2be7c2949bb1b87e52d784971061164d05ecc404318"} Oct 10 13:36:15 crc kubenswrapper[4745]: I1010 13:36:15.631519 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41ddef85-570e-4646-beed-88c7f0492ffc","Type":"ContainerStarted","Data":"1354f0aa8f1618070678aa51f2731227f202cfe0e839ee2053026c3fa16057bb"} Oct 10 13:36:15 crc kubenswrapper[4745]: I1010 13:36:15.631575 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 13:36:15 crc kubenswrapper[4745]: I1010 13:36:15.652847 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-ctcrf" podStartSLOduration=2.652826937 podStartE2EDuration="2.652826937s" podCreationTimestamp="2025-10-10 13:36:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:36:15.640480908 +0000 UTC m=+1089.538137681" watchObservedRunningTime="2025-10-10 13:36:15.652826937 +0000 UTC m=+1089.550483700" Oct 10 13:36:15 crc kubenswrapper[4745]: I1010 13:36:15.678343 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.95047934 podStartE2EDuration="5.678322664s" podCreationTimestamp="2025-10-10 13:36:10 +0000 UTC" firstStartedPulling="2025-10-10 13:36:11.367722876 +0000 UTC m=+1085.265379679" lastFinishedPulling="2025-10-10 13:36:15.09556624 +0000 UTC m=+1088.993223003" observedRunningTime="2025-10-10 13:36:15.668646422 +0000 UTC m=+1089.566303215" watchObservedRunningTime="2025-10-10 13:36:15.678322664 +0000 UTC m=+1089.575979437" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.021000 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.114005 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-cp2b9"] Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.114576 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" podUID="ab0f388b-1766-4cf5-8d43-6811ab7263d5" containerName="dnsmasq-dns" containerID="cri-o://c5828bbc40ad4151ca605349b1eaffcd0f4ad4f1db8fd07fcee8f418d0f5e6e0" gracePeriod=10 Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.187438 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.187531 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.589548 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.671685 4745 generic.go:334] "Generic (PLEG): container finished" podID="ab0f388b-1766-4cf5-8d43-6811ab7263d5" containerID="c5828bbc40ad4151ca605349b1eaffcd0f4ad4f1db8fd07fcee8f418d0f5e6e0" exitCode=0 Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.671786 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" event={"ID":"ab0f388b-1766-4cf5-8d43-6811ab7263d5","Type":"ContainerDied","Data":"c5828bbc40ad4151ca605349b1eaffcd0f4ad4f1db8fd07fcee8f418d0f5e6e0"} Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.671866 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" event={"ID":"ab0f388b-1766-4cf5-8d43-6811ab7263d5","Type":"ContainerDied","Data":"ce5fc0dbfe107bdb6a9d61cf471c35aa4ded16a847cb27d12ab24a02a595c815"} Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.671892 4745 scope.go:117] "RemoveContainer" containerID="c5828bbc40ad4151ca605349b1eaffcd0f4ad4f1db8fd07fcee8f418d0f5e6e0" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.673430 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-cp2b9" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.696740 4745 scope.go:117] "RemoveContainer" containerID="5c29b77ee52a4466794a32b6d88addde506044a430165b2e9100f0de056c466b" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.719119 4745 scope.go:117] "RemoveContainer" containerID="c5828bbc40ad4151ca605349b1eaffcd0f4ad4f1db8fd07fcee8f418d0f5e6e0" Oct 10 13:36:16 crc kubenswrapper[4745]: E1010 13:36:16.719602 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5828bbc40ad4151ca605349b1eaffcd0f4ad4f1db8fd07fcee8f418d0f5e6e0\": container with ID starting with c5828bbc40ad4151ca605349b1eaffcd0f4ad4f1db8fd07fcee8f418d0f5e6e0 not found: ID does not exist" containerID="c5828bbc40ad4151ca605349b1eaffcd0f4ad4f1db8fd07fcee8f418d0f5e6e0" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.719652 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5828bbc40ad4151ca605349b1eaffcd0f4ad4f1db8fd07fcee8f418d0f5e6e0"} err="failed to get container status \"c5828bbc40ad4151ca605349b1eaffcd0f4ad4f1db8fd07fcee8f418d0f5e6e0\": rpc error: code = NotFound desc = could not find container \"c5828bbc40ad4151ca605349b1eaffcd0f4ad4f1db8fd07fcee8f418d0f5e6e0\": container with ID starting with c5828bbc40ad4151ca605349b1eaffcd0f4ad4f1db8fd07fcee8f418d0f5e6e0 not found: ID does not exist" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.719678 4745 scope.go:117] "RemoveContainer" containerID="5c29b77ee52a4466794a32b6d88addde506044a430165b2e9100f0de056c466b" Oct 10 13:36:16 crc kubenswrapper[4745]: E1010 13:36:16.720065 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c29b77ee52a4466794a32b6d88addde506044a430165b2e9100f0de056c466b\": container with ID starting with 5c29b77ee52a4466794a32b6d88addde506044a430165b2e9100f0de056c466b not found: ID does not exist" containerID="5c29b77ee52a4466794a32b6d88addde506044a430165b2e9100f0de056c466b" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.720196 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c29b77ee52a4466794a32b6d88addde506044a430165b2e9100f0de056c466b"} err="failed to get container status \"5c29b77ee52a4466794a32b6d88addde506044a430165b2e9100f0de056c466b\": rpc error: code = NotFound desc = could not find container \"5c29b77ee52a4466794a32b6d88addde506044a430165b2e9100f0de056c466b\": container with ID starting with 5c29b77ee52a4466794a32b6d88addde506044a430165b2e9100f0de056c466b not found: ID does not exist" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.762346 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-ovsdbserver-nb\") pod \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.762441 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srb4n\" (UniqueName: \"kubernetes.io/projected/ab0f388b-1766-4cf5-8d43-6811ab7263d5-kube-api-access-srb4n\") pod \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.762489 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-dns-swift-storage-0\") pod \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.762515 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-config\") pod \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.762629 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-ovsdbserver-sb\") pod \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.762651 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-dns-svc\") pod \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\" (UID: \"ab0f388b-1766-4cf5-8d43-6811ab7263d5\") " Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.768611 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab0f388b-1766-4cf5-8d43-6811ab7263d5-kube-api-access-srb4n" (OuterVolumeSpecName: "kube-api-access-srb4n") pod "ab0f388b-1766-4cf5-8d43-6811ab7263d5" (UID: "ab0f388b-1766-4cf5-8d43-6811ab7263d5"). InnerVolumeSpecName "kube-api-access-srb4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.817815 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ab0f388b-1766-4cf5-8d43-6811ab7263d5" (UID: "ab0f388b-1766-4cf5-8d43-6811ab7263d5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.822717 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ab0f388b-1766-4cf5-8d43-6811ab7263d5" (UID: "ab0f388b-1766-4cf5-8d43-6811ab7263d5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.826458 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ab0f388b-1766-4cf5-8d43-6811ab7263d5" (UID: "ab0f388b-1766-4cf5-8d43-6811ab7263d5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.833712 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-config" (OuterVolumeSpecName: "config") pod "ab0f388b-1766-4cf5-8d43-6811ab7263d5" (UID: "ab0f388b-1766-4cf5-8d43-6811ab7263d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.837361 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ab0f388b-1766-4cf5-8d43-6811ab7263d5" (UID: "ab0f388b-1766-4cf5-8d43-6811ab7263d5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.865528 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.865571 4745 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.865586 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.865601 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srb4n\" (UniqueName: \"kubernetes.io/projected/ab0f388b-1766-4cf5-8d43-6811ab7263d5-kube-api-access-srb4n\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.865613 4745 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:16 crc kubenswrapper[4745]: I1010 13:36:16.865625 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab0f388b-1766-4cf5-8d43-6811ab7263d5-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:17 crc kubenswrapper[4745]: I1010 13:36:17.001953 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-cp2b9"] Oct 10 13:36:17 crc kubenswrapper[4745]: I1010 13:36:17.011818 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-cp2b9"] Oct 10 13:36:18 crc kubenswrapper[4745]: I1010 13:36:18.757490 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab0f388b-1766-4cf5-8d43-6811ab7263d5" path="/var/lib/kubelet/pods/ab0f388b-1766-4cf5-8d43-6811ab7263d5/volumes" Oct 10 13:36:19 crc kubenswrapper[4745]: I1010 13:36:19.710309 4745 generic.go:334] "Generic (PLEG): container finished" podID="16ee733c-3513-40bc-97ec-515a9f735a56" containerID="2e6f2227fb0e20b163ece2be7c2949bb1b87e52d784971061164d05ecc404318" exitCode=0 Oct 10 13:36:19 crc kubenswrapper[4745]: I1010 13:36:19.710520 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ctcrf" event={"ID":"16ee733c-3513-40bc-97ec-515a9f735a56","Type":"ContainerDied","Data":"2e6f2227fb0e20b163ece2be7c2949bb1b87e52d784971061164d05ecc404318"} Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.246600 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ctcrf" Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.358493 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-config-data\") pod \"16ee733c-3513-40bc-97ec-515a9f735a56\" (UID: \"16ee733c-3513-40bc-97ec-515a9f735a56\") " Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.359561 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgcf4\" (UniqueName: \"kubernetes.io/projected/16ee733c-3513-40bc-97ec-515a9f735a56-kube-api-access-pgcf4\") pod \"16ee733c-3513-40bc-97ec-515a9f735a56\" (UID: \"16ee733c-3513-40bc-97ec-515a9f735a56\") " Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.359706 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-combined-ca-bundle\") pod \"16ee733c-3513-40bc-97ec-515a9f735a56\" (UID: \"16ee733c-3513-40bc-97ec-515a9f735a56\") " Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.359961 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-scripts\") pod \"16ee733c-3513-40bc-97ec-515a9f735a56\" (UID: \"16ee733c-3513-40bc-97ec-515a9f735a56\") " Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.364329 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-scripts" (OuterVolumeSpecName: "scripts") pod "16ee733c-3513-40bc-97ec-515a9f735a56" (UID: "16ee733c-3513-40bc-97ec-515a9f735a56"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.371913 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16ee733c-3513-40bc-97ec-515a9f735a56-kube-api-access-pgcf4" (OuterVolumeSpecName: "kube-api-access-pgcf4") pod "16ee733c-3513-40bc-97ec-515a9f735a56" (UID: "16ee733c-3513-40bc-97ec-515a9f735a56"). InnerVolumeSpecName "kube-api-access-pgcf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.408651 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-config-data" (OuterVolumeSpecName: "config-data") pod "16ee733c-3513-40bc-97ec-515a9f735a56" (UID: "16ee733c-3513-40bc-97ec-515a9f735a56"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.410723 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16ee733c-3513-40bc-97ec-515a9f735a56" (UID: "16ee733c-3513-40bc-97ec-515a9f735a56"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.462445 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.462482 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.462495 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ee733c-3513-40bc-97ec-515a9f735a56-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.462507 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgcf4\" (UniqueName: \"kubernetes.io/projected/16ee733c-3513-40bc-97ec-515a9f735a56-kube-api-access-pgcf4\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.747127 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ctcrf" event={"ID":"16ee733c-3513-40bc-97ec-515a9f735a56","Type":"ContainerDied","Data":"36870fbeb7dc4c49dff284ca7cf53df3e52b38ea09feaac6149a58622e83698e"} Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.747175 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36870fbeb7dc4c49dff284ca7cf53df3e52b38ea09feaac6149a58622e83698e" Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.747246 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ctcrf" Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.921080 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.921367 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="085f48ec-754f-4822-a379-b6fb7e4c44b3" containerName="nova-api-log" containerID="cri-o://02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042" gracePeriod=30 Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.921432 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="085f48ec-754f-4822-a379-b6fb7e4c44b3" containerName="nova-api-api" containerID="cri-o://ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96" gracePeriod=30 Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.949536 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.949868 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="29c84d1a-c34f-4331-86c1-7c0f2184d98c" containerName="nova-scheduler-scheduler" containerID="cri-o://5875dfe77b16322bc95114bd035d3465d421d7e7e7904dcb40e1431a18869ff1" gracePeriod=30 Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.978271 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.978539 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0c8be38c-cfc0-4877-a6cb-297a8cc27c73" containerName="nova-metadata-log" containerID="cri-o://ad3fe94cc64220fec72e0955b52deac7a6ee30abed948ce06a3a7ef8cc04eea0" gracePeriod=30 Oct 10 13:36:21 crc kubenswrapper[4745]: I1010 13:36:21.978638 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0c8be38c-cfc0-4877-a6cb-297a8cc27c73" containerName="nova-metadata-metadata" containerID="cri-o://90bc1ee4e8020dc1894d4e1c0c175b564baa6d96f3f9d719e9e0e6b85a06e86c" gracePeriod=30 Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.500295 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.585106 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/085f48ec-754f-4822-a379-b6fb7e4c44b3-logs\") pod \"085f48ec-754f-4822-a379-b6fb7e4c44b3\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.585147 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-config-data\") pod \"085f48ec-754f-4822-a379-b6fb7e4c44b3\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.585187 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-internal-tls-certs\") pod \"085f48ec-754f-4822-a379-b6fb7e4c44b3\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.585223 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-combined-ca-bundle\") pod \"085f48ec-754f-4822-a379-b6fb7e4c44b3\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.585247 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-public-tls-certs\") pod \"085f48ec-754f-4822-a379-b6fb7e4c44b3\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.585283 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rspmz\" (UniqueName: \"kubernetes.io/projected/085f48ec-754f-4822-a379-b6fb7e4c44b3-kube-api-access-rspmz\") pod \"085f48ec-754f-4822-a379-b6fb7e4c44b3\" (UID: \"085f48ec-754f-4822-a379-b6fb7e4c44b3\") " Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.585463 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/085f48ec-754f-4822-a379-b6fb7e4c44b3-logs" (OuterVolumeSpecName: "logs") pod "085f48ec-754f-4822-a379-b6fb7e4c44b3" (UID: "085f48ec-754f-4822-a379-b6fb7e4c44b3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.585762 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/085f48ec-754f-4822-a379-b6fb7e4c44b3-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.591843 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/085f48ec-754f-4822-a379-b6fb7e4c44b3-kube-api-access-rspmz" (OuterVolumeSpecName: "kube-api-access-rspmz") pod "085f48ec-754f-4822-a379-b6fb7e4c44b3" (UID: "085f48ec-754f-4822-a379-b6fb7e4c44b3"). InnerVolumeSpecName "kube-api-access-rspmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.618337 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-config-data" (OuterVolumeSpecName: "config-data") pod "085f48ec-754f-4822-a379-b6fb7e4c44b3" (UID: "085f48ec-754f-4822-a379-b6fb7e4c44b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.631410 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "085f48ec-754f-4822-a379-b6fb7e4c44b3" (UID: "085f48ec-754f-4822-a379-b6fb7e4c44b3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.644867 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "085f48ec-754f-4822-a379-b6fb7e4c44b3" (UID: "085f48ec-754f-4822-a379-b6fb7e4c44b3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.653827 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "085f48ec-754f-4822-a379-b6fb7e4c44b3" (UID: "085f48ec-754f-4822-a379-b6fb7e4c44b3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.686873 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.686900 4745 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.686911 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.686921 4745 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/085f48ec-754f-4822-a379-b6fb7e4c44b3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.686932 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rspmz\" (UniqueName: \"kubernetes.io/projected/085f48ec-754f-4822-a379-b6fb7e4c44b3-kube-api-access-rspmz\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.758721 4745 generic.go:334] "Generic (PLEG): container finished" podID="0c8be38c-cfc0-4877-a6cb-297a8cc27c73" containerID="ad3fe94cc64220fec72e0955b52deac7a6ee30abed948ce06a3a7ef8cc04eea0" exitCode=143 Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.758769 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0c8be38c-cfc0-4877-a6cb-297a8cc27c73","Type":"ContainerDied","Data":"ad3fe94cc64220fec72e0955b52deac7a6ee30abed948ce06a3a7ef8cc04eea0"} Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.760979 4745 generic.go:334] "Generic (PLEG): container finished" podID="085f48ec-754f-4822-a379-b6fb7e4c44b3" containerID="ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96" exitCode=0 Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.761002 4745 generic.go:334] "Generic (PLEG): container finished" podID="085f48ec-754f-4822-a379-b6fb7e4c44b3" containerID="02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042" exitCode=143 Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.761008 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"085f48ec-754f-4822-a379-b6fb7e4c44b3","Type":"ContainerDied","Data":"ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96"} Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.761038 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"085f48ec-754f-4822-a379-b6fb7e4c44b3","Type":"ContainerDied","Data":"02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042"} Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.761054 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"085f48ec-754f-4822-a379-b6fb7e4c44b3","Type":"ContainerDied","Data":"ffab2c4096dd0039e5b1a0164c981aa75dd24e98b88a632176cd77e53329c1e1"} Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.761070 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.761075 4745 scope.go:117] "RemoveContainer" containerID="ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.780636 4745 scope.go:117] "RemoveContainer" containerID="02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.793574 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.806974 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.814636 4745 scope.go:117] "RemoveContainer" containerID="ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96" Oct 10 13:36:22 crc kubenswrapper[4745]: E1010 13:36:22.815052 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96\": container with ID starting with ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96 not found: ID does not exist" containerID="ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.815089 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96"} err="failed to get container status \"ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96\": rpc error: code = NotFound desc = could not find container \"ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96\": container with ID starting with ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96 not found: ID does not exist" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.815116 4745 scope.go:117] "RemoveContainer" containerID="02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042" Oct 10 13:36:22 crc kubenswrapper[4745]: E1010 13:36:22.815359 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042\": container with ID starting with 02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042 not found: ID does not exist" containerID="02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.815388 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042"} err="failed to get container status \"02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042\": rpc error: code = NotFound desc = could not find container \"02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042\": container with ID starting with 02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042 not found: ID does not exist" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.815407 4745 scope.go:117] "RemoveContainer" containerID="ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.815853 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96"} err="failed to get container status \"ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96\": rpc error: code = NotFound desc = could not find container \"ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96\": container with ID starting with ea1d2ce0bbe59eee4a7c90e640c0fc1d11a8269d31f65b72c1d9bdea79378a96 not found: ID does not exist" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.815874 4745 scope.go:117] "RemoveContainer" containerID="02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.816171 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042"} err="failed to get container status \"02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042\": rpc error: code = NotFound desc = could not find container \"02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042\": container with ID starting with 02373246f20956ca15e3d73a62afe4de8848fd984532f04759b807358bda3042 not found: ID does not exist" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.821872 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 10 13:36:22 crc kubenswrapper[4745]: E1010 13:36:22.822219 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab0f388b-1766-4cf5-8d43-6811ab7263d5" containerName="init" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.822237 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab0f388b-1766-4cf5-8d43-6811ab7263d5" containerName="init" Oct 10 13:36:22 crc kubenswrapper[4745]: E1010 13:36:22.822261 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="085f48ec-754f-4822-a379-b6fb7e4c44b3" containerName="nova-api-api" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.822267 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="085f48ec-754f-4822-a379-b6fb7e4c44b3" containerName="nova-api-api" Oct 10 13:36:22 crc kubenswrapper[4745]: E1010 13:36:22.822280 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab0f388b-1766-4cf5-8d43-6811ab7263d5" containerName="dnsmasq-dns" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.822286 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab0f388b-1766-4cf5-8d43-6811ab7263d5" containerName="dnsmasq-dns" Oct 10 13:36:22 crc kubenswrapper[4745]: E1010 13:36:22.822300 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16ee733c-3513-40bc-97ec-515a9f735a56" containerName="nova-manage" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.822306 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="16ee733c-3513-40bc-97ec-515a9f735a56" containerName="nova-manage" Oct 10 13:36:22 crc kubenswrapper[4745]: E1010 13:36:22.822321 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="085f48ec-754f-4822-a379-b6fb7e4c44b3" containerName="nova-api-log" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.822327 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="085f48ec-754f-4822-a379-b6fb7e4c44b3" containerName="nova-api-log" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.822517 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="085f48ec-754f-4822-a379-b6fb7e4c44b3" containerName="nova-api-api" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.822529 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="16ee733c-3513-40bc-97ec-515a9f735a56" containerName="nova-manage" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.822539 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab0f388b-1766-4cf5-8d43-6811ab7263d5" containerName="dnsmasq-dns" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.822557 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="085f48ec-754f-4822-a379-b6fb7e4c44b3" containerName="nova-api-log" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.823696 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.825763 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.826308 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.826494 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.830259 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.991990 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1801b5b-c008-463f-952b-92ddef21faae-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.992323 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6rdm\" (UniqueName: \"kubernetes.io/projected/b1801b5b-c008-463f-952b-92ddef21faae-kube-api-access-m6rdm\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.992386 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1801b5b-c008-463f-952b-92ddef21faae-config-data\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.992601 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1801b5b-c008-463f-952b-92ddef21faae-public-tls-certs\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.992773 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1801b5b-c008-463f-952b-92ddef21faae-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:22 crc kubenswrapper[4745]: I1010 13:36:22.992829 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1801b5b-c008-463f-952b-92ddef21faae-logs\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.094377 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6rdm\" (UniqueName: \"kubernetes.io/projected/b1801b5b-c008-463f-952b-92ddef21faae-kube-api-access-m6rdm\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.094419 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1801b5b-c008-463f-952b-92ddef21faae-config-data\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.094451 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1801b5b-c008-463f-952b-92ddef21faae-public-tls-certs\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.094485 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1801b5b-c008-463f-952b-92ddef21faae-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.094505 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1801b5b-c008-463f-952b-92ddef21faae-logs\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.094531 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1801b5b-c008-463f-952b-92ddef21faae-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.095111 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1801b5b-c008-463f-952b-92ddef21faae-logs\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.098584 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1801b5b-c008-463f-952b-92ddef21faae-public-tls-certs\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.098773 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1801b5b-c008-463f-952b-92ddef21faae-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.099212 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1801b5b-c008-463f-952b-92ddef21faae-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.100199 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1801b5b-c008-463f-952b-92ddef21faae-config-data\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.119284 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6rdm\" (UniqueName: \"kubernetes.io/projected/b1801b5b-c008-463f-952b-92ddef21faae-kube-api-access-m6rdm\") pod \"nova-api-0\" (UID: \"b1801b5b-c008-463f-952b-92ddef21faae\") " pod="openstack/nova-api-0" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.139518 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.605856 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 13:36:23 crc kubenswrapper[4745]: W1010 13:36:23.607880 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1801b5b_c008_463f_952b_92ddef21faae.slice/crio-49cbebd4c9dc38841c15e3c652d43badb3588e88b4d6fa5c7939b9e1a3058578 WatchSource:0}: Error finding container 49cbebd4c9dc38841c15e3c652d43badb3588e88b4d6fa5c7939b9e1a3058578: Status 404 returned error can't find the container with id 49cbebd4c9dc38841c15e3c652d43badb3588e88b4d6fa5c7939b9e1a3058578 Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.743111 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.775028 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b1801b5b-c008-463f-952b-92ddef21faae","Type":"ContainerStarted","Data":"49cbebd4c9dc38841c15e3c652d43badb3588e88b4d6fa5c7939b9e1a3058578"} Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.775928 4745 generic.go:334] "Generic (PLEG): container finished" podID="29c84d1a-c34f-4331-86c1-7c0f2184d98c" containerID="5875dfe77b16322bc95114bd035d3465d421d7e7e7904dcb40e1431a18869ff1" exitCode=0 Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.775953 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"29c84d1a-c34f-4331-86c1-7c0f2184d98c","Type":"ContainerDied","Data":"5875dfe77b16322bc95114bd035d3465d421d7e7e7904dcb40e1431a18869ff1"} Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.775969 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"29c84d1a-c34f-4331-86c1-7c0f2184d98c","Type":"ContainerDied","Data":"56fde75b08cd25d7ae37f5af25a77b3ac8a497834a8afb06a8978b9d77750b52"} Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.775984 4745 scope.go:117] "RemoveContainer" containerID="5875dfe77b16322bc95114bd035d3465d421d7e7e7904dcb40e1431a18869ff1" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.776067 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.798765 4745 scope.go:117] "RemoveContainer" containerID="5875dfe77b16322bc95114bd035d3465d421d7e7e7904dcb40e1431a18869ff1" Oct 10 13:36:23 crc kubenswrapper[4745]: E1010 13:36:23.799449 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5875dfe77b16322bc95114bd035d3465d421d7e7e7904dcb40e1431a18869ff1\": container with ID starting with 5875dfe77b16322bc95114bd035d3465d421d7e7e7904dcb40e1431a18869ff1 not found: ID does not exist" containerID="5875dfe77b16322bc95114bd035d3465d421d7e7e7904dcb40e1431a18869ff1" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.799494 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5875dfe77b16322bc95114bd035d3465d421d7e7e7904dcb40e1431a18869ff1"} err="failed to get container status \"5875dfe77b16322bc95114bd035d3465d421d7e7e7904dcb40e1431a18869ff1\": rpc error: code = NotFound desc = could not find container \"5875dfe77b16322bc95114bd035d3465d421d7e7e7904dcb40e1431a18869ff1\": container with ID starting with 5875dfe77b16322bc95114bd035d3465d421d7e7e7904dcb40e1431a18869ff1 not found: ID does not exist" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.909905 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgjrz\" (UniqueName: \"kubernetes.io/projected/29c84d1a-c34f-4331-86c1-7c0f2184d98c-kube-api-access-zgjrz\") pod \"29c84d1a-c34f-4331-86c1-7c0f2184d98c\" (UID: \"29c84d1a-c34f-4331-86c1-7c0f2184d98c\") " Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.910587 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c84d1a-c34f-4331-86c1-7c0f2184d98c-combined-ca-bundle\") pod \"29c84d1a-c34f-4331-86c1-7c0f2184d98c\" (UID: \"29c84d1a-c34f-4331-86c1-7c0f2184d98c\") " Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.910766 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29c84d1a-c34f-4331-86c1-7c0f2184d98c-config-data\") pod \"29c84d1a-c34f-4331-86c1-7c0f2184d98c\" (UID: \"29c84d1a-c34f-4331-86c1-7c0f2184d98c\") " Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.922421 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29c84d1a-c34f-4331-86c1-7c0f2184d98c-kube-api-access-zgjrz" (OuterVolumeSpecName: "kube-api-access-zgjrz") pod "29c84d1a-c34f-4331-86c1-7c0f2184d98c" (UID: "29c84d1a-c34f-4331-86c1-7c0f2184d98c"). InnerVolumeSpecName "kube-api-access-zgjrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.938891 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29c84d1a-c34f-4331-86c1-7c0f2184d98c-config-data" (OuterVolumeSpecName: "config-data") pod "29c84d1a-c34f-4331-86c1-7c0f2184d98c" (UID: "29c84d1a-c34f-4331-86c1-7c0f2184d98c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:23 crc kubenswrapper[4745]: I1010 13:36:23.951277 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29c84d1a-c34f-4331-86c1-7c0f2184d98c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29c84d1a-c34f-4331-86c1-7c0f2184d98c" (UID: "29c84d1a-c34f-4331-86c1-7c0f2184d98c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.013493 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgjrz\" (UniqueName: \"kubernetes.io/projected/29c84d1a-c34f-4331-86c1-7c0f2184d98c-kube-api-access-zgjrz\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.013892 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c84d1a-c34f-4331-86c1-7c0f2184d98c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.013990 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29c84d1a-c34f-4331-86c1-7c0f2184d98c-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.119154 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.131994 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.141465 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 13:36:24 crc kubenswrapper[4745]: E1010 13:36:24.142108 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29c84d1a-c34f-4331-86c1-7c0f2184d98c" containerName="nova-scheduler-scheduler" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.142135 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="29c84d1a-c34f-4331-86c1-7c0f2184d98c" containerName="nova-scheduler-scheduler" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.142480 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="29c84d1a-c34f-4331-86c1-7c0f2184d98c" containerName="nova-scheduler-scheduler" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.143638 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.146441 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.154334 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.217835 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d13d787-633b-4ae3-8fbd-013c3de88709-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8d13d787-633b-4ae3-8fbd-013c3de88709\") " pod="openstack/nova-scheduler-0" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.218036 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d13d787-633b-4ae3-8fbd-013c3de88709-config-data\") pod \"nova-scheduler-0\" (UID: \"8d13d787-633b-4ae3-8fbd-013c3de88709\") " pod="openstack/nova-scheduler-0" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.218116 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4jzg\" (UniqueName: \"kubernetes.io/projected/8d13d787-633b-4ae3-8fbd-013c3de88709-kube-api-access-n4jzg\") pod \"nova-scheduler-0\" (UID: \"8d13d787-633b-4ae3-8fbd-013c3de88709\") " pod="openstack/nova-scheduler-0" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.319548 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d13d787-633b-4ae3-8fbd-013c3de88709-config-data\") pod \"nova-scheduler-0\" (UID: \"8d13d787-633b-4ae3-8fbd-013c3de88709\") " pod="openstack/nova-scheduler-0" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.319942 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4jzg\" (UniqueName: \"kubernetes.io/projected/8d13d787-633b-4ae3-8fbd-013c3de88709-kube-api-access-n4jzg\") pod \"nova-scheduler-0\" (UID: \"8d13d787-633b-4ae3-8fbd-013c3de88709\") " pod="openstack/nova-scheduler-0" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.320007 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d13d787-633b-4ae3-8fbd-013c3de88709-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8d13d787-633b-4ae3-8fbd-013c3de88709\") " pod="openstack/nova-scheduler-0" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.323287 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d13d787-633b-4ae3-8fbd-013c3de88709-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8d13d787-633b-4ae3-8fbd-013c3de88709\") " pod="openstack/nova-scheduler-0" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.328310 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d13d787-633b-4ae3-8fbd-013c3de88709-config-data\") pod \"nova-scheduler-0\" (UID: \"8d13d787-633b-4ae3-8fbd-013c3de88709\") " pod="openstack/nova-scheduler-0" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.346991 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4jzg\" (UniqueName: \"kubernetes.io/projected/8d13d787-633b-4ae3-8fbd-013c3de88709-kube-api-access-n4jzg\") pod \"nova-scheduler-0\" (UID: \"8d13d787-633b-4ae3-8fbd-013c3de88709\") " pod="openstack/nova-scheduler-0" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.508119 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.775061 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="085f48ec-754f-4822-a379-b6fb7e4c44b3" path="/var/lib/kubelet/pods/085f48ec-754f-4822-a379-b6fb7e4c44b3/volumes" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.775957 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29c84d1a-c34f-4331-86c1-7c0f2184d98c" path="/var/lib/kubelet/pods/29c84d1a-c34f-4331-86c1-7c0f2184d98c/volumes" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.787467 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b1801b5b-c008-463f-952b-92ddef21faae","Type":"ContainerStarted","Data":"a16296f20439360857b2b1e3e35642d9c9c2e1dc8f2fd184c8452b1f38621a99"} Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.787519 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b1801b5b-c008-463f-952b-92ddef21faae","Type":"ContainerStarted","Data":"12ef549f83e0d4ff0dc29d8493cd253a602d8c656b181e34a84ac4f7badac377"} Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.810071 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.810055078 podStartE2EDuration="2.810055078s" podCreationTimestamp="2025-10-10 13:36:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:36:24.808310114 +0000 UTC m=+1098.705966887" watchObservedRunningTime="2025-10-10 13:36:24.810055078 +0000 UTC m=+1098.707711841" Oct 10 13:36:24 crc kubenswrapper[4745]: I1010 13:36:24.975509 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.137086 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="0c8be38c-cfc0-4877-a6cb-297a8cc27c73" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:53962->10.217.0.191:8775: read: connection reset by peer" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.137418 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="0c8be38c-cfc0-4877-a6cb-297a8cc27c73" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:53964->10.217.0.191:8775: read: connection reset by peer" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.591325 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.748232 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-logs\") pod \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.748537 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-config-data\") pod \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.748670 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-combined-ca-bundle\") pod \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.748747 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-logs" (OuterVolumeSpecName: "logs") pod "0c8be38c-cfc0-4877-a6cb-297a8cc27c73" (UID: "0c8be38c-cfc0-4877-a6cb-297a8cc27c73"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.749359 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxgc8\" (UniqueName: \"kubernetes.io/projected/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-kube-api-access-qxgc8\") pod \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.749488 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-nova-metadata-tls-certs\") pod \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\" (UID: \"0c8be38c-cfc0-4877-a6cb-297a8cc27c73\") " Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.749995 4745 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-logs\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.755978 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-kube-api-access-qxgc8" (OuterVolumeSpecName: "kube-api-access-qxgc8") pod "0c8be38c-cfc0-4877-a6cb-297a8cc27c73" (UID: "0c8be38c-cfc0-4877-a6cb-297a8cc27c73"). InnerVolumeSpecName "kube-api-access-qxgc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.780408 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c8be38c-cfc0-4877-a6cb-297a8cc27c73" (UID: "0c8be38c-cfc0-4877-a6cb-297a8cc27c73"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.796026 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-config-data" (OuterVolumeSpecName: "config-data") pod "0c8be38c-cfc0-4877-a6cb-297a8cc27c73" (UID: "0c8be38c-cfc0-4877-a6cb-297a8cc27c73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.800248 4745 generic.go:334] "Generic (PLEG): container finished" podID="0c8be38c-cfc0-4877-a6cb-297a8cc27c73" containerID="90bc1ee4e8020dc1894d4e1c0c175b564baa6d96f3f9d719e9e0e6b85a06e86c" exitCode=0 Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.800308 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0c8be38c-cfc0-4877-a6cb-297a8cc27c73","Type":"ContainerDied","Data":"90bc1ee4e8020dc1894d4e1c0c175b564baa6d96f3f9d719e9e0e6b85a06e86c"} Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.800334 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0c8be38c-cfc0-4877-a6cb-297a8cc27c73","Type":"ContainerDied","Data":"52f7a5169ea38de2643de6b330578618baced73141f976fa1c418d5981ffe250"} Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.800350 4745 scope.go:117] "RemoveContainer" containerID="90bc1ee4e8020dc1894d4e1c0c175b564baa6d96f3f9d719e9e0e6b85a06e86c" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.800461 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.804821 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8d13d787-633b-4ae3-8fbd-013c3de88709","Type":"ContainerStarted","Data":"5c457e285f0ea9c142b515491d7c44d7a9b2a90b96231d5b70a8d2f5b02c1cf2"} Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.804874 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8d13d787-633b-4ae3-8fbd-013c3de88709","Type":"ContainerStarted","Data":"7dea09e0dddf03bab2a3f3af011f6bfcecbbc9ccdfd9d195c844077ac4926390"} Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.810904 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "0c8be38c-cfc0-4877-a6cb-297a8cc27c73" (UID: "0c8be38c-cfc0-4877-a6cb-297a8cc27c73"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.831177 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.831154637 podStartE2EDuration="1.831154637s" podCreationTimestamp="2025-10-10 13:36:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:36:25.826913581 +0000 UTC m=+1099.724570354" watchObservedRunningTime="2025-10-10 13:36:25.831154637 +0000 UTC m=+1099.728811410" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.852380 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.852424 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxgc8\" (UniqueName: \"kubernetes.io/projected/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-kube-api-access-qxgc8\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.852434 4745 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.852443 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c8be38c-cfc0-4877-a6cb-297a8cc27c73-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.857441 4745 scope.go:117] "RemoveContainer" containerID="ad3fe94cc64220fec72e0955b52deac7a6ee30abed948ce06a3a7ef8cc04eea0" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.883328 4745 scope.go:117] "RemoveContainer" containerID="90bc1ee4e8020dc1894d4e1c0c175b564baa6d96f3f9d719e9e0e6b85a06e86c" Oct 10 13:36:25 crc kubenswrapper[4745]: E1010 13:36:25.883902 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90bc1ee4e8020dc1894d4e1c0c175b564baa6d96f3f9d719e9e0e6b85a06e86c\": container with ID starting with 90bc1ee4e8020dc1894d4e1c0c175b564baa6d96f3f9d719e9e0e6b85a06e86c not found: ID does not exist" containerID="90bc1ee4e8020dc1894d4e1c0c175b564baa6d96f3f9d719e9e0e6b85a06e86c" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.883942 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90bc1ee4e8020dc1894d4e1c0c175b564baa6d96f3f9d719e9e0e6b85a06e86c"} err="failed to get container status \"90bc1ee4e8020dc1894d4e1c0c175b564baa6d96f3f9d719e9e0e6b85a06e86c\": rpc error: code = NotFound desc = could not find container \"90bc1ee4e8020dc1894d4e1c0c175b564baa6d96f3f9d719e9e0e6b85a06e86c\": container with ID starting with 90bc1ee4e8020dc1894d4e1c0c175b564baa6d96f3f9d719e9e0e6b85a06e86c not found: ID does not exist" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.883966 4745 scope.go:117] "RemoveContainer" containerID="ad3fe94cc64220fec72e0955b52deac7a6ee30abed948ce06a3a7ef8cc04eea0" Oct 10 13:36:25 crc kubenswrapper[4745]: E1010 13:36:25.884154 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad3fe94cc64220fec72e0955b52deac7a6ee30abed948ce06a3a7ef8cc04eea0\": container with ID starting with ad3fe94cc64220fec72e0955b52deac7a6ee30abed948ce06a3a7ef8cc04eea0 not found: ID does not exist" containerID="ad3fe94cc64220fec72e0955b52deac7a6ee30abed948ce06a3a7ef8cc04eea0" Oct 10 13:36:25 crc kubenswrapper[4745]: I1010 13:36:25.884178 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad3fe94cc64220fec72e0955b52deac7a6ee30abed948ce06a3a7ef8cc04eea0"} err="failed to get container status \"ad3fe94cc64220fec72e0955b52deac7a6ee30abed948ce06a3a7ef8cc04eea0\": rpc error: code = NotFound desc = could not find container \"ad3fe94cc64220fec72e0955b52deac7a6ee30abed948ce06a3a7ef8cc04eea0\": container with ID starting with ad3fe94cc64220fec72e0955b52deac7a6ee30abed948ce06a3a7ef8cc04eea0 not found: ID does not exist" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.142240 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.159778 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.180983 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:36:26 crc kubenswrapper[4745]: E1010 13:36:26.182281 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c8be38c-cfc0-4877-a6cb-297a8cc27c73" containerName="nova-metadata-metadata" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.182313 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c8be38c-cfc0-4877-a6cb-297a8cc27c73" containerName="nova-metadata-metadata" Oct 10 13:36:26 crc kubenswrapper[4745]: E1010 13:36:26.182378 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c8be38c-cfc0-4877-a6cb-297a8cc27c73" containerName="nova-metadata-log" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.182387 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c8be38c-cfc0-4877-a6cb-297a8cc27c73" containerName="nova-metadata-log" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.182884 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c8be38c-cfc0-4877-a6cb-297a8cc27c73" containerName="nova-metadata-metadata" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.182955 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c8be38c-cfc0-4877-a6cb-297a8cc27c73" containerName="nova-metadata-log" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.185251 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.188196 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.188591 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.199592 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.263515 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/254ea8d2-b80e-40c0-8e33-3459e73878de-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"254ea8d2-b80e-40c0-8e33-3459e73878de\") " pod="openstack/nova-metadata-0" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.263560 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/254ea8d2-b80e-40c0-8e33-3459e73878de-config-data\") pod \"nova-metadata-0\" (UID: \"254ea8d2-b80e-40c0-8e33-3459e73878de\") " pod="openstack/nova-metadata-0" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.263623 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjsh8\" (UniqueName: \"kubernetes.io/projected/254ea8d2-b80e-40c0-8e33-3459e73878de-kube-api-access-sjsh8\") pod \"nova-metadata-0\" (UID: \"254ea8d2-b80e-40c0-8e33-3459e73878de\") " pod="openstack/nova-metadata-0" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.263639 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/254ea8d2-b80e-40c0-8e33-3459e73878de-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"254ea8d2-b80e-40c0-8e33-3459e73878de\") " pod="openstack/nova-metadata-0" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.263664 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/254ea8d2-b80e-40c0-8e33-3459e73878de-logs\") pod \"nova-metadata-0\" (UID: \"254ea8d2-b80e-40c0-8e33-3459e73878de\") " pod="openstack/nova-metadata-0" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.365257 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/254ea8d2-b80e-40c0-8e33-3459e73878de-logs\") pod \"nova-metadata-0\" (UID: \"254ea8d2-b80e-40c0-8e33-3459e73878de\") " pod="openstack/nova-metadata-0" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.365611 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/254ea8d2-b80e-40c0-8e33-3459e73878de-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"254ea8d2-b80e-40c0-8e33-3459e73878de\") " pod="openstack/nova-metadata-0" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.365666 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/254ea8d2-b80e-40c0-8e33-3459e73878de-logs\") pod \"nova-metadata-0\" (UID: \"254ea8d2-b80e-40c0-8e33-3459e73878de\") " pod="openstack/nova-metadata-0" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.365758 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/254ea8d2-b80e-40c0-8e33-3459e73878de-config-data\") pod \"nova-metadata-0\" (UID: \"254ea8d2-b80e-40c0-8e33-3459e73878de\") " pod="openstack/nova-metadata-0" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.366012 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjsh8\" (UniqueName: \"kubernetes.io/projected/254ea8d2-b80e-40c0-8e33-3459e73878de-kube-api-access-sjsh8\") pod \"nova-metadata-0\" (UID: \"254ea8d2-b80e-40c0-8e33-3459e73878de\") " pod="openstack/nova-metadata-0" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.366120 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/254ea8d2-b80e-40c0-8e33-3459e73878de-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"254ea8d2-b80e-40c0-8e33-3459e73878de\") " pod="openstack/nova-metadata-0" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.369266 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/254ea8d2-b80e-40c0-8e33-3459e73878de-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"254ea8d2-b80e-40c0-8e33-3459e73878de\") " pod="openstack/nova-metadata-0" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.369960 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/254ea8d2-b80e-40c0-8e33-3459e73878de-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"254ea8d2-b80e-40c0-8e33-3459e73878de\") " pod="openstack/nova-metadata-0" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.376294 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/254ea8d2-b80e-40c0-8e33-3459e73878de-config-data\") pod \"nova-metadata-0\" (UID: \"254ea8d2-b80e-40c0-8e33-3459e73878de\") " pod="openstack/nova-metadata-0" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.384039 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjsh8\" (UniqueName: \"kubernetes.io/projected/254ea8d2-b80e-40c0-8e33-3459e73878de-kube-api-access-sjsh8\") pod \"nova-metadata-0\" (UID: \"254ea8d2-b80e-40c0-8e33-3459e73878de\") " pod="openstack/nova-metadata-0" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.553455 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 13:36:26 crc kubenswrapper[4745]: I1010 13:36:26.771547 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c8be38c-cfc0-4877-a6cb-297a8cc27c73" path="/var/lib/kubelet/pods/0c8be38c-cfc0-4877-a6cb-297a8cc27c73/volumes" Oct 10 13:36:27 crc kubenswrapper[4745]: I1010 13:36:27.017856 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 13:36:27 crc kubenswrapper[4745]: W1010 13:36:27.022134 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod254ea8d2_b80e_40c0_8e33_3459e73878de.slice/crio-29a3337b96bb4f4e7a6e041c2eee0bddc5428d47eb807b0f9b6b3834acc8ed16 WatchSource:0}: Error finding container 29a3337b96bb4f4e7a6e041c2eee0bddc5428d47eb807b0f9b6b3834acc8ed16: Status 404 returned error can't find the container with id 29a3337b96bb4f4e7a6e041c2eee0bddc5428d47eb807b0f9b6b3834acc8ed16 Oct 10 13:36:27 crc kubenswrapper[4745]: I1010 13:36:27.828146 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"254ea8d2-b80e-40c0-8e33-3459e73878de","Type":"ContainerStarted","Data":"fd097cdfece82bd70b1422b0c8d57a42d02d463b09acfd04f76953cba04b7ad2"} Oct 10 13:36:27 crc kubenswrapper[4745]: I1010 13:36:27.828491 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"254ea8d2-b80e-40c0-8e33-3459e73878de","Type":"ContainerStarted","Data":"1afc1265e0d99da4030a9ce3791da0112e2e52de73f9547063c67b6034a15f85"} Oct 10 13:36:27 crc kubenswrapper[4745]: I1010 13:36:27.828501 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"254ea8d2-b80e-40c0-8e33-3459e73878de","Type":"ContainerStarted","Data":"29a3337b96bb4f4e7a6e041c2eee0bddc5428d47eb807b0f9b6b3834acc8ed16"} Oct 10 13:36:27 crc kubenswrapper[4745]: I1010 13:36:27.865715 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.865682934 podStartE2EDuration="1.865682934s" podCreationTimestamp="2025-10-10 13:36:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:36:27.848711903 +0000 UTC m=+1101.746368696" watchObservedRunningTime="2025-10-10 13:36:27.865682934 +0000 UTC m=+1101.763339737" Oct 10 13:36:29 crc kubenswrapper[4745]: I1010 13:36:29.508834 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 10 13:36:31 crc kubenswrapper[4745]: I1010 13:36:31.554321 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 13:36:31 crc kubenswrapper[4745]: I1010 13:36:31.554409 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 13:36:33 crc kubenswrapper[4745]: I1010 13:36:33.140269 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 13:36:33 crc kubenswrapper[4745]: I1010 13:36:33.140939 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 13:36:34 crc kubenswrapper[4745]: I1010 13:36:34.161959 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b1801b5b-c008-463f-952b-92ddef21faae" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 13:36:34 crc kubenswrapper[4745]: I1010 13:36:34.162013 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b1801b5b-c008-463f-952b-92ddef21faae" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 13:36:34 crc kubenswrapper[4745]: I1010 13:36:34.509012 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 10 13:36:34 crc kubenswrapper[4745]: I1010 13:36:34.541843 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 10 13:36:34 crc kubenswrapper[4745]: I1010 13:36:34.952430 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 10 13:36:36 crc kubenswrapper[4745]: I1010 13:36:36.553907 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 13:36:36 crc kubenswrapper[4745]: I1010 13:36:36.554239 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 13:36:37 crc kubenswrapper[4745]: I1010 13:36:37.567121 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="254ea8d2-b80e-40c0-8e33-3459e73878de" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 13:36:37 crc kubenswrapper[4745]: I1010 13:36:37.567181 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="254ea8d2-b80e-40c0-8e33-3459e73878de" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 13:36:40 crc kubenswrapper[4745]: I1010 13:36:40.896995 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 10 13:36:43 crc kubenswrapper[4745]: I1010 13:36:43.147437 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 13:36:43 crc kubenswrapper[4745]: I1010 13:36:43.148768 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 13:36:43 crc kubenswrapper[4745]: I1010 13:36:43.149542 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 13:36:43 crc kubenswrapper[4745]: I1010 13:36:43.154791 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 13:36:44 crc kubenswrapper[4745]: I1010 13:36:44.026175 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 13:36:44 crc kubenswrapper[4745]: I1010 13:36:44.033201 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 13:36:44 crc kubenswrapper[4745]: I1010 13:36:44.818387 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 13:36:44 crc kubenswrapper[4745]: I1010 13:36:44.818973 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="2f59de38-158a-4c6d-b902-daed00239a27" containerName="kube-state-metrics" containerID="cri-o://7220f11f38ff2c790ce1a10f2004b3a7b290380db4339fcad096fc5bbe639915" gracePeriod=30 Oct 10 13:36:45 crc kubenswrapper[4745]: I1010 13:36:45.042575 4745 generic.go:334] "Generic (PLEG): container finished" podID="2f59de38-158a-4c6d-b902-daed00239a27" containerID="7220f11f38ff2c790ce1a10f2004b3a7b290380db4339fcad096fc5bbe639915" exitCode=2 Oct 10 13:36:45 crc kubenswrapper[4745]: I1010 13:36:45.043052 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2f59de38-158a-4c6d-b902-daed00239a27","Type":"ContainerDied","Data":"7220f11f38ff2c790ce1a10f2004b3a7b290380db4339fcad096fc5bbe639915"} Oct 10 13:36:45 crc kubenswrapper[4745]: I1010 13:36:45.344110 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 13:36:45 crc kubenswrapper[4745]: I1010 13:36:45.438453 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4pgr\" (UniqueName: \"kubernetes.io/projected/2f59de38-158a-4c6d-b902-daed00239a27-kube-api-access-b4pgr\") pod \"2f59de38-158a-4c6d-b902-daed00239a27\" (UID: \"2f59de38-158a-4c6d-b902-daed00239a27\") " Oct 10 13:36:45 crc kubenswrapper[4745]: I1010 13:36:45.448898 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f59de38-158a-4c6d-b902-daed00239a27-kube-api-access-b4pgr" (OuterVolumeSpecName: "kube-api-access-b4pgr") pod "2f59de38-158a-4c6d-b902-daed00239a27" (UID: "2f59de38-158a-4c6d-b902-daed00239a27"). InnerVolumeSpecName "kube-api-access-b4pgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:36:45 crc kubenswrapper[4745]: I1010 13:36:45.540201 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4pgr\" (UniqueName: \"kubernetes.io/projected/2f59de38-158a-4c6d-b902-daed00239a27-kube-api-access-b4pgr\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.059059 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2f59de38-158a-4c6d-b902-daed00239a27","Type":"ContainerDied","Data":"cccdd845bf25820d8e10ffd92184d358a1b2c2603d53ab7ae17dca89bf6a6f26"} Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.059082 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.059391 4745 scope.go:117] "RemoveContainer" containerID="7220f11f38ff2c790ce1a10f2004b3a7b290380db4339fcad096fc5bbe639915" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.094868 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.106960 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.157039 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 13:36:46 crc kubenswrapper[4745]: E1010 13:36:46.159316 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f59de38-158a-4c6d-b902-daed00239a27" containerName="kube-state-metrics" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.159360 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f59de38-158a-4c6d-b902-daed00239a27" containerName="kube-state-metrics" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.159917 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f59de38-158a-4c6d-b902-daed00239a27" containerName="kube-state-metrics" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.163906 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.166696 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.166765 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.174826 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.186470 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.186528 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.186572 4745 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.187359 4745 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"051be97974334031c215babb8b965fd2290d7b1b439e8dabb7659be3724a636f"} pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.187428 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" containerID="cri-o://051be97974334031c215babb8b965fd2290d7b1b439e8dabb7659be3724a636f" gracePeriod=600 Oct 10 13:36:46 crc kubenswrapper[4745]: E1010 13:36:46.230197 4745 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f59de38_158a_4c6d_b902_daed00239a27.slice\": RecentStats: unable to find data in memory cache]" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.260808 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c1adab6-47ba-4e16-b607-e61f92db1b53-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"2c1adab6-47ba-4e16-b607-e61f92db1b53\") " pod="openstack/kube-state-metrics-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.260907 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/2c1adab6-47ba-4e16-b607-e61f92db1b53-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"2c1adab6-47ba-4e16-b607-e61f92db1b53\") " pod="openstack/kube-state-metrics-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.260936 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1adab6-47ba-4e16-b607-e61f92db1b53-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"2c1adab6-47ba-4e16-b607-e61f92db1b53\") " pod="openstack/kube-state-metrics-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.260968 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glxx4\" (UniqueName: \"kubernetes.io/projected/2c1adab6-47ba-4e16-b607-e61f92db1b53-kube-api-access-glxx4\") pod \"kube-state-metrics-0\" (UID: \"2c1adab6-47ba-4e16-b607-e61f92db1b53\") " pod="openstack/kube-state-metrics-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.362713 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/2c1adab6-47ba-4e16-b607-e61f92db1b53-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"2c1adab6-47ba-4e16-b607-e61f92db1b53\") " pod="openstack/kube-state-metrics-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.363134 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1adab6-47ba-4e16-b607-e61f92db1b53-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"2c1adab6-47ba-4e16-b607-e61f92db1b53\") " pod="openstack/kube-state-metrics-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.363188 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glxx4\" (UniqueName: \"kubernetes.io/projected/2c1adab6-47ba-4e16-b607-e61f92db1b53-kube-api-access-glxx4\") pod \"kube-state-metrics-0\" (UID: \"2c1adab6-47ba-4e16-b607-e61f92db1b53\") " pod="openstack/kube-state-metrics-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.363324 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c1adab6-47ba-4e16-b607-e61f92db1b53-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"2c1adab6-47ba-4e16-b607-e61f92db1b53\") " pod="openstack/kube-state-metrics-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.370208 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c1adab6-47ba-4e16-b607-e61f92db1b53-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"2c1adab6-47ba-4e16-b607-e61f92db1b53\") " pod="openstack/kube-state-metrics-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.370389 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1adab6-47ba-4e16-b607-e61f92db1b53-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"2c1adab6-47ba-4e16-b607-e61f92db1b53\") " pod="openstack/kube-state-metrics-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.371894 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/2c1adab6-47ba-4e16-b607-e61f92db1b53-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"2c1adab6-47ba-4e16-b607-e61f92db1b53\") " pod="openstack/kube-state-metrics-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.380886 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glxx4\" (UniqueName: \"kubernetes.io/projected/2c1adab6-47ba-4e16-b607-e61f92db1b53-kube-api-access-glxx4\") pod \"kube-state-metrics-0\" (UID: \"2c1adab6-47ba-4e16-b607-e61f92db1b53\") " pod="openstack/kube-state-metrics-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.491172 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.546064 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.546447 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41ddef85-570e-4646-beed-88c7f0492ffc" containerName="ceilometer-central-agent" containerID="cri-o://57eafa639f92ba605f2bdef31e303ca544d927866b6c441756fe1ab099939887" gracePeriod=30 Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.546587 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41ddef85-570e-4646-beed-88c7f0492ffc" containerName="proxy-httpd" containerID="cri-o://1354f0aa8f1618070678aa51f2731227f202cfe0e839ee2053026c3fa16057bb" gracePeriod=30 Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.546626 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41ddef85-570e-4646-beed-88c7f0492ffc" containerName="sg-core" containerID="cri-o://38d579f086d9c930d1b9219cb7e9cdbaa805b2fc2db70b85854c7b6f5ff62d14" gracePeriod=30 Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.546668 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41ddef85-570e-4646-beed-88c7f0492ffc" containerName="ceilometer-notification-agent" containerID="cri-o://03d239f2c573a14d66f379afaf5947b1ef2c2196749ecf9c506ce139bd438b94" gracePeriod=30 Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.573091 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.574538 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.582060 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 13:36:46 crc kubenswrapper[4745]: I1010 13:36:46.758674 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f59de38-158a-4c6d-b902-daed00239a27" path="/var/lib/kubelet/pods/2f59de38-158a-4c6d-b902-daed00239a27/volumes" Oct 10 13:36:47 crc kubenswrapper[4745]: I1010 13:36:47.005637 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 13:36:47 crc kubenswrapper[4745]: W1010 13:36:47.008490 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c1adab6_47ba_4e16_b607_e61f92db1b53.slice/crio-0e15ce922659052baa7276be4e6babe8ed69adb742c854f3c8176ca00797a69b WatchSource:0}: Error finding container 0e15ce922659052baa7276be4e6babe8ed69adb742c854f3c8176ca00797a69b: Status 404 returned error can't find the container with id 0e15ce922659052baa7276be4e6babe8ed69adb742c854f3c8176ca00797a69b Oct 10 13:36:47 crc kubenswrapper[4745]: I1010 13:36:47.070190 4745 generic.go:334] "Generic (PLEG): container finished" podID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerID="051be97974334031c215babb8b965fd2290d7b1b439e8dabb7659be3724a636f" exitCode=0 Oct 10 13:36:47 crc kubenswrapper[4745]: I1010 13:36:47.070265 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerDied","Data":"051be97974334031c215babb8b965fd2290d7b1b439e8dabb7659be3724a636f"} Oct 10 13:36:47 crc kubenswrapper[4745]: I1010 13:36:47.070313 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"38f61d2405362274ab0410a7304f61bdd266d1575d542c8d7a802125d8ce76c0"} Oct 10 13:36:47 crc kubenswrapper[4745]: I1010 13:36:47.070335 4745 scope.go:117] "RemoveContainer" containerID="8a31c6076bf04e3e0b289d6ba4cb6b6d1eeb3e2ba416958cc06b859417161f1e" Oct 10 13:36:47 crc kubenswrapper[4745]: I1010 13:36:47.072413 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2c1adab6-47ba-4e16-b607-e61f92db1b53","Type":"ContainerStarted","Data":"0e15ce922659052baa7276be4e6babe8ed69adb742c854f3c8176ca00797a69b"} Oct 10 13:36:47 crc kubenswrapper[4745]: I1010 13:36:47.076747 4745 generic.go:334] "Generic (PLEG): container finished" podID="41ddef85-570e-4646-beed-88c7f0492ffc" containerID="1354f0aa8f1618070678aa51f2731227f202cfe0e839ee2053026c3fa16057bb" exitCode=0 Oct 10 13:36:47 crc kubenswrapper[4745]: I1010 13:36:47.076779 4745 generic.go:334] "Generic (PLEG): container finished" podID="41ddef85-570e-4646-beed-88c7f0492ffc" containerID="38d579f086d9c930d1b9219cb7e9cdbaa805b2fc2db70b85854c7b6f5ff62d14" exitCode=2 Oct 10 13:36:47 crc kubenswrapper[4745]: I1010 13:36:47.076792 4745 generic.go:334] "Generic (PLEG): container finished" podID="41ddef85-570e-4646-beed-88c7f0492ffc" containerID="57eafa639f92ba605f2bdef31e303ca544d927866b6c441756fe1ab099939887" exitCode=0 Oct 10 13:36:47 crc kubenswrapper[4745]: I1010 13:36:47.078794 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41ddef85-570e-4646-beed-88c7f0492ffc","Type":"ContainerDied","Data":"1354f0aa8f1618070678aa51f2731227f202cfe0e839ee2053026c3fa16057bb"} Oct 10 13:36:47 crc kubenswrapper[4745]: I1010 13:36:47.078834 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41ddef85-570e-4646-beed-88c7f0492ffc","Type":"ContainerDied","Data":"38d579f086d9c930d1b9219cb7e9cdbaa805b2fc2db70b85854c7b6f5ff62d14"} Oct 10 13:36:47 crc kubenswrapper[4745]: I1010 13:36:47.078855 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41ddef85-570e-4646-beed-88c7f0492ffc","Type":"ContainerDied","Data":"57eafa639f92ba605f2bdef31e303ca544d927866b6c441756fe1ab099939887"} Oct 10 13:36:47 crc kubenswrapper[4745]: I1010 13:36:47.090881 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.093408 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2c1adab6-47ba-4e16-b607-e61f92db1b53","Type":"ContainerStarted","Data":"487fac14d29f3de166b47291da450022b7e125d9d70ff8a359c6436f4f597e14"} Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.094272 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.121792 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.778635527 podStartE2EDuration="2.121763764s" podCreationTimestamp="2025-10-10 13:36:46 +0000 UTC" firstStartedPulling="2025-10-10 13:36:47.01081799 +0000 UTC m=+1120.908474753" lastFinishedPulling="2025-10-10 13:36:47.353946227 +0000 UTC m=+1121.251602990" observedRunningTime="2025-10-10 13:36:48.113108314 +0000 UTC m=+1122.010765117" watchObservedRunningTime="2025-10-10 13:36:48.121763764 +0000 UTC m=+1122.019420567" Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.837945 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.925660 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-config-data\") pod \"41ddef85-570e-4646-beed-88c7f0492ffc\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.925768 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-scripts\") pod \"41ddef85-570e-4646-beed-88c7f0492ffc\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.925894 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41ddef85-570e-4646-beed-88c7f0492ffc-log-httpd\") pod \"41ddef85-570e-4646-beed-88c7f0492ffc\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.925949 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-combined-ca-bundle\") pod \"41ddef85-570e-4646-beed-88c7f0492ffc\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.926073 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-sg-core-conf-yaml\") pod \"41ddef85-570e-4646-beed-88c7f0492ffc\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.926100 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fswlz\" (UniqueName: \"kubernetes.io/projected/41ddef85-570e-4646-beed-88c7f0492ffc-kube-api-access-fswlz\") pod \"41ddef85-570e-4646-beed-88c7f0492ffc\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.926151 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41ddef85-570e-4646-beed-88c7f0492ffc-run-httpd\") pod \"41ddef85-570e-4646-beed-88c7f0492ffc\" (UID: \"41ddef85-570e-4646-beed-88c7f0492ffc\") " Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.926555 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41ddef85-570e-4646-beed-88c7f0492ffc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "41ddef85-570e-4646-beed-88c7f0492ffc" (UID: "41ddef85-570e-4646-beed-88c7f0492ffc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.926594 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41ddef85-570e-4646-beed-88c7f0492ffc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "41ddef85-570e-4646-beed-88c7f0492ffc" (UID: "41ddef85-570e-4646-beed-88c7f0492ffc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.927025 4745 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41ddef85-570e-4646-beed-88c7f0492ffc-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.927050 4745 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41ddef85-570e-4646-beed-88c7f0492ffc-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.945433 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41ddef85-570e-4646-beed-88c7f0492ffc-kube-api-access-fswlz" (OuterVolumeSpecName: "kube-api-access-fswlz") pod "41ddef85-570e-4646-beed-88c7f0492ffc" (UID: "41ddef85-570e-4646-beed-88c7f0492ffc"). InnerVolumeSpecName "kube-api-access-fswlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.947566 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-scripts" (OuterVolumeSpecName: "scripts") pod "41ddef85-570e-4646-beed-88c7f0492ffc" (UID: "41ddef85-570e-4646-beed-88c7f0492ffc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:48 crc kubenswrapper[4745]: I1010 13:36:48.965927 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "41ddef85-570e-4646-beed-88c7f0492ffc" (UID: "41ddef85-570e-4646-beed-88c7f0492ffc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.010744 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41ddef85-570e-4646-beed-88c7f0492ffc" (UID: "41ddef85-570e-4646-beed-88c7f0492ffc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.029666 4745 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.029890 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.029987 4745 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.030050 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fswlz\" (UniqueName: \"kubernetes.io/projected/41ddef85-570e-4646-beed-88c7f0492ffc-kube-api-access-fswlz\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.042236 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-config-data" (OuterVolumeSpecName: "config-data") pod "41ddef85-570e-4646-beed-88c7f0492ffc" (UID: "41ddef85-570e-4646-beed-88c7f0492ffc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.105634 4745 generic.go:334] "Generic (PLEG): container finished" podID="41ddef85-570e-4646-beed-88c7f0492ffc" containerID="03d239f2c573a14d66f379afaf5947b1ef2c2196749ecf9c506ce139bd438b94" exitCode=0 Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.105684 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.105713 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41ddef85-570e-4646-beed-88c7f0492ffc","Type":"ContainerDied","Data":"03d239f2c573a14d66f379afaf5947b1ef2c2196749ecf9c506ce139bd438b94"} Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.106014 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41ddef85-570e-4646-beed-88c7f0492ffc","Type":"ContainerDied","Data":"2110f4339ed875a72e5059cb55143b049e92787837626654fdc4c5595d5e833f"} Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.106035 4745 scope.go:117] "RemoveContainer" containerID="1354f0aa8f1618070678aa51f2731227f202cfe0e839ee2053026c3fa16057bb" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.133292 4745 scope.go:117] "RemoveContainer" containerID="38d579f086d9c930d1b9219cb7e9cdbaa805b2fc2db70b85854c7b6f5ff62d14" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.134426 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41ddef85-570e-4646-beed-88c7f0492ffc-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.141335 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.147623 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.164169 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:36:49 crc kubenswrapper[4745]: E1010 13:36:49.164523 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ddef85-570e-4646-beed-88c7f0492ffc" containerName="sg-core" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.164539 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ddef85-570e-4646-beed-88c7f0492ffc" containerName="sg-core" Oct 10 13:36:49 crc kubenswrapper[4745]: E1010 13:36:49.164573 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ddef85-570e-4646-beed-88c7f0492ffc" containerName="proxy-httpd" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.164579 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ddef85-570e-4646-beed-88c7f0492ffc" containerName="proxy-httpd" Oct 10 13:36:49 crc kubenswrapper[4745]: E1010 13:36:49.164592 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ddef85-570e-4646-beed-88c7f0492ffc" containerName="ceilometer-central-agent" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.164598 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ddef85-570e-4646-beed-88c7f0492ffc" containerName="ceilometer-central-agent" Oct 10 13:36:49 crc kubenswrapper[4745]: E1010 13:36:49.164615 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ddef85-570e-4646-beed-88c7f0492ffc" containerName="ceilometer-notification-agent" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.164620 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ddef85-570e-4646-beed-88c7f0492ffc" containerName="ceilometer-notification-agent" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.164795 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ddef85-570e-4646-beed-88c7f0492ffc" containerName="ceilometer-central-agent" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.164807 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ddef85-570e-4646-beed-88c7f0492ffc" containerName="sg-core" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.164820 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ddef85-570e-4646-beed-88c7f0492ffc" containerName="proxy-httpd" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.164836 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ddef85-570e-4646-beed-88c7f0492ffc" containerName="ceilometer-notification-agent" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.167949 4745 scope.go:117] "RemoveContainer" containerID="03d239f2c573a14d66f379afaf5947b1ef2c2196749ecf9c506ce139bd438b94" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.174189 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.177595 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.177811 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.178573 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.179598 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.186625 4745 scope.go:117] "RemoveContainer" containerID="57eafa639f92ba605f2bdef31e303ca544d927866b6c441756fe1ab099939887" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.214433 4745 scope.go:117] "RemoveContainer" containerID="1354f0aa8f1618070678aa51f2731227f202cfe0e839ee2053026c3fa16057bb" Oct 10 13:36:49 crc kubenswrapper[4745]: E1010 13:36:49.214919 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1354f0aa8f1618070678aa51f2731227f202cfe0e839ee2053026c3fa16057bb\": container with ID starting with 1354f0aa8f1618070678aa51f2731227f202cfe0e839ee2053026c3fa16057bb not found: ID does not exist" containerID="1354f0aa8f1618070678aa51f2731227f202cfe0e839ee2053026c3fa16057bb" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.215030 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1354f0aa8f1618070678aa51f2731227f202cfe0e839ee2053026c3fa16057bb"} err="failed to get container status \"1354f0aa8f1618070678aa51f2731227f202cfe0e839ee2053026c3fa16057bb\": rpc error: code = NotFound desc = could not find container \"1354f0aa8f1618070678aa51f2731227f202cfe0e839ee2053026c3fa16057bb\": container with ID starting with 1354f0aa8f1618070678aa51f2731227f202cfe0e839ee2053026c3fa16057bb not found: ID does not exist" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.215112 4745 scope.go:117] "RemoveContainer" containerID="38d579f086d9c930d1b9219cb7e9cdbaa805b2fc2db70b85854c7b6f5ff62d14" Oct 10 13:36:49 crc kubenswrapper[4745]: E1010 13:36:49.215590 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38d579f086d9c930d1b9219cb7e9cdbaa805b2fc2db70b85854c7b6f5ff62d14\": container with ID starting with 38d579f086d9c930d1b9219cb7e9cdbaa805b2fc2db70b85854c7b6f5ff62d14 not found: ID does not exist" containerID="38d579f086d9c930d1b9219cb7e9cdbaa805b2fc2db70b85854c7b6f5ff62d14" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.215629 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38d579f086d9c930d1b9219cb7e9cdbaa805b2fc2db70b85854c7b6f5ff62d14"} err="failed to get container status \"38d579f086d9c930d1b9219cb7e9cdbaa805b2fc2db70b85854c7b6f5ff62d14\": rpc error: code = NotFound desc = could not find container \"38d579f086d9c930d1b9219cb7e9cdbaa805b2fc2db70b85854c7b6f5ff62d14\": container with ID starting with 38d579f086d9c930d1b9219cb7e9cdbaa805b2fc2db70b85854c7b6f5ff62d14 not found: ID does not exist" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.215655 4745 scope.go:117] "RemoveContainer" containerID="03d239f2c573a14d66f379afaf5947b1ef2c2196749ecf9c506ce139bd438b94" Oct 10 13:36:49 crc kubenswrapper[4745]: E1010 13:36:49.215960 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03d239f2c573a14d66f379afaf5947b1ef2c2196749ecf9c506ce139bd438b94\": container with ID starting with 03d239f2c573a14d66f379afaf5947b1ef2c2196749ecf9c506ce139bd438b94 not found: ID does not exist" containerID="03d239f2c573a14d66f379afaf5947b1ef2c2196749ecf9c506ce139bd438b94" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.215997 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03d239f2c573a14d66f379afaf5947b1ef2c2196749ecf9c506ce139bd438b94"} err="failed to get container status \"03d239f2c573a14d66f379afaf5947b1ef2c2196749ecf9c506ce139bd438b94\": rpc error: code = NotFound desc = could not find container \"03d239f2c573a14d66f379afaf5947b1ef2c2196749ecf9c506ce139bd438b94\": container with ID starting with 03d239f2c573a14d66f379afaf5947b1ef2c2196749ecf9c506ce139bd438b94 not found: ID does not exist" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.216012 4745 scope.go:117] "RemoveContainer" containerID="57eafa639f92ba605f2bdef31e303ca544d927866b6c441756fe1ab099939887" Oct 10 13:36:49 crc kubenswrapper[4745]: E1010 13:36:49.216275 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57eafa639f92ba605f2bdef31e303ca544d927866b6c441756fe1ab099939887\": container with ID starting with 57eafa639f92ba605f2bdef31e303ca544d927866b6c441756fe1ab099939887 not found: ID does not exist" containerID="57eafa639f92ba605f2bdef31e303ca544d927866b6c441756fe1ab099939887" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.216300 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57eafa639f92ba605f2bdef31e303ca544d927866b6c441756fe1ab099939887"} err="failed to get container status \"57eafa639f92ba605f2bdef31e303ca544d927866b6c441756fe1ab099939887\": rpc error: code = NotFound desc = could not find container \"57eafa639f92ba605f2bdef31e303ca544d927866b6c441756fe1ab099939887\": container with ID starting with 57eafa639f92ba605f2bdef31e303ca544d927866b6c441756fe1ab099939887 not found: ID does not exist" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.235608 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d791e6b-724b-4631-8a50-98b6dfd108f2-config-data\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.235671 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d791e6b-724b-4631-8a50-98b6dfd108f2-log-httpd\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.235762 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d791e6b-724b-4631-8a50-98b6dfd108f2-scripts\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.235786 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d791e6b-724b-4631-8a50-98b6dfd108f2-run-httpd\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.235825 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d791e6b-724b-4631-8a50-98b6dfd108f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.235933 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d791e6b-724b-4631-8a50-98b6dfd108f2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.235990 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf8d7\" (UniqueName: \"kubernetes.io/projected/0d791e6b-724b-4631-8a50-98b6dfd108f2-kube-api-access-nf8d7\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.236029 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d791e6b-724b-4631-8a50-98b6dfd108f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.338423 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d791e6b-724b-4631-8a50-98b6dfd108f2-scripts\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.338491 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d791e6b-724b-4631-8a50-98b6dfd108f2-run-httpd\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.338548 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d791e6b-724b-4631-8a50-98b6dfd108f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.338658 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d791e6b-724b-4631-8a50-98b6dfd108f2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.338707 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf8d7\" (UniqueName: \"kubernetes.io/projected/0d791e6b-724b-4631-8a50-98b6dfd108f2-kube-api-access-nf8d7\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.338742 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d791e6b-724b-4631-8a50-98b6dfd108f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.338797 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d791e6b-724b-4631-8a50-98b6dfd108f2-config-data\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.338826 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d791e6b-724b-4631-8a50-98b6dfd108f2-log-httpd\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.339082 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d791e6b-724b-4631-8a50-98b6dfd108f2-log-httpd\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.339930 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d791e6b-724b-4631-8a50-98b6dfd108f2-run-httpd\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.343804 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d791e6b-724b-4631-8a50-98b6dfd108f2-scripts\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.344106 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d791e6b-724b-4631-8a50-98b6dfd108f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.344130 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d791e6b-724b-4631-8a50-98b6dfd108f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.344227 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d791e6b-724b-4631-8a50-98b6dfd108f2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.344603 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d791e6b-724b-4631-8a50-98b6dfd108f2-config-data\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.354165 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf8d7\" (UniqueName: \"kubernetes.io/projected/0d791e6b-724b-4631-8a50-98b6dfd108f2-kube-api-access-nf8d7\") pod \"ceilometer-0\" (UID: \"0d791e6b-724b-4631-8a50-98b6dfd108f2\") " pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.495536 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 13:36:49 crc kubenswrapper[4745]: W1010 13:36:49.991289 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d791e6b_724b_4631_8a50_98b6dfd108f2.slice/crio-29d1bc2d21f9467029f3e55164b94922e2e4fdb8af5a10719a8e3739b28aa655 WatchSource:0}: Error finding container 29d1bc2d21f9467029f3e55164b94922e2e4fdb8af5a10719a8e3739b28aa655: Status 404 returned error can't find the container with id 29d1bc2d21f9467029f3e55164b94922e2e4fdb8af5a10719a8e3739b28aa655 Oct 10 13:36:49 crc kubenswrapper[4745]: I1010 13:36:49.993478 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 13:36:50 crc kubenswrapper[4745]: I1010 13:36:50.120718 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d791e6b-724b-4631-8a50-98b6dfd108f2","Type":"ContainerStarted","Data":"29d1bc2d21f9467029f3e55164b94922e2e4fdb8af5a10719a8e3739b28aa655"} Oct 10 13:36:50 crc kubenswrapper[4745]: I1010 13:36:50.755842 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41ddef85-570e-4646-beed-88c7f0492ffc" path="/var/lib/kubelet/pods/41ddef85-570e-4646-beed-88c7f0492ffc/volumes" Oct 10 13:36:51 crc kubenswrapper[4745]: I1010 13:36:51.143306 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d791e6b-724b-4631-8a50-98b6dfd108f2","Type":"ContainerStarted","Data":"ca94280ac21bd4b2ca3101c1ea6c716447467c158904507f2dc7bd949f2d6a61"} Oct 10 13:36:52 crc kubenswrapper[4745]: I1010 13:36:52.157314 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d791e6b-724b-4631-8a50-98b6dfd108f2","Type":"ContainerStarted","Data":"7262c3244e923779f94005bd130549f01bf54e0c7bea77310f43cfb704648e86"} Oct 10 13:36:53 crc kubenswrapper[4745]: I1010 13:36:53.170126 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d791e6b-724b-4631-8a50-98b6dfd108f2","Type":"ContainerStarted","Data":"f55d44720904e6035b1c8db1e49674ae9aa3cd8717fde52b2415501d8f808eb1"} Oct 10 13:36:54 crc kubenswrapper[4745]: I1010 13:36:54.183881 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d791e6b-724b-4631-8a50-98b6dfd108f2","Type":"ContainerStarted","Data":"1c92b1a1658a3d670d4926a39a60787dfa2af49892842180c9471d4153218b0a"} Oct 10 13:36:54 crc kubenswrapper[4745]: I1010 13:36:54.184495 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 13:36:54 crc kubenswrapper[4745]: I1010 13:36:54.223800 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.423350826 podStartE2EDuration="5.223779206s" podCreationTimestamp="2025-10-10 13:36:49 +0000 UTC" firstStartedPulling="2025-10-10 13:36:49.993254889 +0000 UTC m=+1123.890911672" lastFinishedPulling="2025-10-10 13:36:53.793683289 +0000 UTC m=+1127.691340052" observedRunningTime="2025-10-10 13:36:54.215842864 +0000 UTC m=+1128.113499667" watchObservedRunningTime="2025-10-10 13:36:54.223779206 +0000 UTC m=+1128.121435969" Oct 10 13:36:56 crc kubenswrapper[4745]: I1010 13:36:56.506826 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 10 13:37:19 crc kubenswrapper[4745]: I1010 13:37:19.516452 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 10 13:37:29 crc kubenswrapper[4745]: I1010 13:37:29.051446 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 13:37:30 crc kubenswrapper[4745]: I1010 13:37:30.500861 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 13:37:32 crc kubenswrapper[4745]: I1010 13:37:32.946457 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="bc0bb94e-36a7-413d-9ce0-eea2031d1d19" containerName="rabbitmq" containerID="cri-o://83f39747e46356d5c08ec33d69e30ab2569a1f7461903ed89ce9864a0017be49" gracePeriod=604797 Oct 10 13:37:34 crc kubenswrapper[4745]: I1010 13:37:34.250785 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="234155a2-9751-459f-b259-6bb32eb4b4d7" containerName="rabbitmq" containerID="cri-o://c7ea1a57fde8811da41997a7bcd0d2d8f76592e019f75dd145aad49f9c8962d6" gracePeriod=604797 Oct 10 13:37:34 crc kubenswrapper[4745]: I1010 13:37:34.963017 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="bc0bb94e-36a7-413d-9ce0-eea2031d1d19" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Oct 10 13:37:35 crc kubenswrapper[4745]: I1010 13:37:35.234781 4745 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="234155a2-9751-459f-b259-6bb32eb4b4d7" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.512081 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.676105 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-config-data\") pod \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.676375 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-confd\") pod \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.676445 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5g5m\" (UniqueName: \"kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-kube-api-access-r5g5m\") pod \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.676467 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.676515 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-plugins\") pod \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.676544 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-tls\") pod \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.676596 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-plugins-conf\") pod \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.676623 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-erlang-cookie\") pod \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.676639 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-erlang-cookie-secret\") pod \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.676671 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-server-conf\") pod \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.676692 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-pod-info\") pod \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\" (UID: \"bc0bb94e-36a7-413d-9ce0-eea2031d1d19\") " Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.677365 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "bc0bb94e-36a7-413d-9ce0-eea2031d1d19" (UID: "bc0bb94e-36a7-413d-9ce0-eea2031d1d19"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.677416 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "bc0bb94e-36a7-413d-9ce0-eea2031d1d19" (UID: "bc0bb94e-36a7-413d-9ce0-eea2031d1d19"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.677866 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "bc0bb94e-36a7-413d-9ce0-eea2031d1d19" (UID: "bc0bb94e-36a7-413d-9ce0-eea2031d1d19"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.681928 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-pod-info" (OuterVolumeSpecName: "pod-info") pod "bc0bb94e-36a7-413d-9ce0-eea2031d1d19" (UID: "bc0bb94e-36a7-413d-9ce0-eea2031d1d19"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.682169 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "bc0bb94e-36a7-413d-9ce0-eea2031d1d19" (UID: "bc0bb94e-36a7-413d-9ce0-eea2031d1d19"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.682884 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "bc0bb94e-36a7-413d-9ce0-eea2031d1d19" (UID: "bc0bb94e-36a7-413d-9ce0-eea2031d1d19"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.684599 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-kube-api-access-r5g5m" (OuterVolumeSpecName: "kube-api-access-r5g5m") pod "bc0bb94e-36a7-413d-9ce0-eea2031d1d19" (UID: "bc0bb94e-36a7-413d-9ce0-eea2031d1d19"). InnerVolumeSpecName "kube-api-access-r5g5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.695102 4745 generic.go:334] "Generic (PLEG): container finished" podID="bc0bb94e-36a7-413d-9ce0-eea2031d1d19" containerID="83f39747e46356d5c08ec33d69e30ab2569a1f7461903ed89ce9864a0017be49" exitCode=0 Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.695163 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.695187 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bc0bb94e-36a7-413d-9ce0-eea2031d1d19","Type":"ContainerDied","Data":"83f39747e46356d5c08ec33d69e30ab2569a1f7461903ed89ce9864a0017be49"} Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.695218 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bc0bb94e-36a7-413d-9ce0-eea2031d1d19","Type":"ContainerDied","Data":"26e4f12fcb056d774ac4b8b7f3ca3349a8a2a9165229875488fd6158fc9fdfe6"} Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.695233 4745 scope.go:117] "RemoveContainer" containerID="83f39747e46356d5c08ec33d69e30ab2569a1f7461903ed89ce9864a0017be49" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.696640 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "bc0bb94e-36a7-413d-9ce0-eea2031d1d19" (UID: "bc0bb94e-36a7-413d-9ce0-eea2031d1d19"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.735939 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-config-data" (OuterVolumeSpecName: "config-data") pod "bc0bb94e-36a7-413d-9ce0-eea2031d1d19" (UID: "bc0bb94e-36a7-413d-9ce0-eea2031d1d19"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.758150 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-server-conf" (OuterVolumeSpecName: "server-conf") pod "bc0bb94e-36a7-413d-9ce0-eea2031d1d19" (UID: "bc0bb94e-36a7-413d-9ce0-eea2031d1d19"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.778732 4745 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.778779 4745 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.778790 4745 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.778803 4745 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-server-conf\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.778811 4745 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-pod-info\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.778819 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.778827 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5g5m\" (UniqueName: \"kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-kube-api-access-r5g5m\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.778845 4745 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.778853 4745 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.778861 4745 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.792939 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "bc0bb94e-36a7-413d-9ce0-eea2031d1d19" (UID: "bc0bb94e-36a7-413d-9ce0-eea2031d1d19"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.807140 4745 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.814446 4745 scope.go:117] "RemoveContainer" containerID="e9b2c0742bac99a38637cce65f207a0590e5ea6ba54dc0c7bcb8a39f75955347" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.842082 4745 scope.go:117] "RemoveContainer" containerID="83f39747e46356d5c08ec33d69e30ab2569a1f7461903ed89ce9864a0017be49" Oct 10 13:37:39 crc kubenswrapper[4745]: E1010 13:37:39.842791 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83f39747e46356d5c08ec33d69e30ab2569a1f7461903ed89ce9864a0017be49\": container with ID starting with 83f39747e46356d5c08ec33d69e30ab2569a1f7461903ed89ce9864a0017be49 not found: ID does not exist" containerID="83f39747e46356d5c08ec33d69e30ab2569a1f7461903ed89ce9864a0017be49" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.842828 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83f39747e46356d5c08ec33d69e30ab2569a1f7461903ed89ce9864a0017be49"} err="failed to get container status \"83f39747e46356d5c08ec33d69e30ab2569a1f7461903ed89ce9864a0017be49\": rpc error: code = NotFound desc = could not find container \"83f39747e46356d5c08ec33d69e30ab2569a1f7461903ed89ce9864a0017be49\": container with ID starting with 83f39747e46356d5c08ec33d69e30ab2569a1f7461903ed89ce9864a0017be49 not found: ID does not exist" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.842854 4745 scope.go:117] "RemoveContainer" containerID="e9b2c0742bac99a38637cce65f207a0590e5ea6ba54dc0c7bcb8a39f75955347" Oct 10 13:37:39 crc kubenswrapper[4745]: E1010 13:37:39.843209 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9b2c0742bac99a38637cce65f207a0590e5ea6ba54dc0c7bcb8a39f75955347\": container with ID starting with e9b2c0742bac99a38637cce65f207a0590e5ea6ba54dc0c7bcb8a39f75955347 not found: ID does not exist" containerID="e9b2c0742bac99a38637cce65f207a0590e5ea6ba54dc0c7bcb8a39f75955347" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.843251 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9b2c0742bac99a38637cce65f207a0590e5ea6ba54dc0c7bcb8a39f75955347"} err="failed to get container status \"e9b2c0742bac99a38637cce65f207a0590e5ea6ba54dc0c7bcb8a39f75955347\": rpc error: code = NotFound desc = could not find container \"e9b2c0742bac99a38637cce65f207a0590e5ea6ba54dc0c7bcb8a39f75955347\": container with ID starting with e9b2c0742bac99a38637cce65f207a0590e5ea6ba54dc0c7bcb8a39f75955347 not found: ID does not exist" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.879792 4745 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc0bb94e-36a7-413d-9ce0-eea2031d1d19-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:39 crc kubenswrapper[4745]: I1010 13:37:39.879827 4745 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.031431 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.044210 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.058122 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 13:37:40 crc kubenswrapper[4745]: E1010 13:37:40.058646 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc0bb94e-36a7-413d-9ce0-eea2031d1d19" containerName="rabbitmq" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.058672 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc0bb94e-36a7-413d-9ce0-eea2031d1d19" containerName="rabbitmq" Oct 10 13:37:40 crc kubenswrapper[4745]: E1010 13:37:40.058723 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc0bb94e-36a7-413d-9ce0-eea2031d1d19" containerName="setup-container" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.058749 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc0bb94e-36a7-413d-9ce0-eea2031d1d19" containerName="setup-container" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.059010 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc0bb94e-36a7-413d-9ce0-eea2031d1d19" containerName="rabbitmq" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.060283 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.063320 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.063642 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.063882 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.064042 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.064193 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.064315 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-tkvfq" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.064420 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.074951 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.184194 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-config-data\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.184278 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.184379 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.184419 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.184437 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.184685 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.184743 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.184783 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.184798 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.184838 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.184903 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c52nl\" (UniqueName: \"kubernetes.io/projected/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-kube-api-access-c52nl\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.286042 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-config-data\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.286099 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.286120 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.286135 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.286149 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.286224 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.286244 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.286266 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.286279 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.286299 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.286322 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c52nl\" (UniqueName: \"kubernetes.io/projected/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-kube-api-access-c52nl\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.286977 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.287196 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-config-data\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.287581 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.287670 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.287891 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.287990 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.291141 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.291301 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.294682 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.298212 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.302940 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c52nl\" (UniqueName: \"kubernetes.io/projected/3f7d0aba-ccad-4d20-9ab1-cf7d080571c1-kube-api-access-c52nl\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.339471 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1\") " pod="openstack/rabbitmq-server-0" Oct 10 13:37:40 crc kubenswrapper[4745]: I1010 13:37:40.392004 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.713787 4745 generic.go:334] "Generic (PLEG): container finished" podID="234155a2-9751-459f-b259-6bb32eb4b4d7" containerID="c7ea1a57fde8811da41997a7bcd0d2d8f76592e019f75dd145aad49f9c8962d6" exitCode=0 Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.714003 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"234155a2-9751-459f-b259-6bb32eb4b4d7","Type":"ContainerDied","Data":"c7ea1a57fde8811da41997a7bcd0d2d8f76592e019f75dd145aad49f9c8962d6"} Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.755154 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc0bb94e-36a7-413d-9ce0-eea2031d1d19" path="/var/lib/kubelet/pods/bc0bb94e-36a7-413d-9ce0-eea2031d1d19/volumes" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.788229 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.882139 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.898811 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-confd\") pod \"234155a2-9751-459f-b259-6bb32eb4b4d7\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.899164 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-plugins-conf\") pod \"234155a2-9751-459f-b259-6bb32eb4b4d7\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.899241 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-erlang-cookie\") pod \"234155a2-9751-459f-b259-6bb32eb4b4d7\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.899292 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/234155a2-9751-459f-b259-6bb32eb4b4d7-pod-info\") pod \"234155a2-9751-459f-b259-6bb32eb4b4d7\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.899436 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-tls\") pod \"234155a2-9751-459f-b259-6bb32eb4b4d7\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.899464 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-config-data\") pod \"234155a2-9751-459f-b259-6bb32eb4b4d7\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.899495 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44b7l\" (UniqueName: \"kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-kube-api-access-44b7l\") pod \"234155a2-9751-459f-b259-6bb32eb4b4d7\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.899567 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/234155a2-9751-459f-b259-6bb32eb4b4d7-erlang-cookie-secret\") pod \"234155a2-9751-459f-b259-6bb32eb4b4d7\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.899618 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"234155a2-9751-459f-b259-6bb32eb4b4d7\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.899646 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-plugins\") pod \"234155a2-9751-459f-b259-6bb32eb4b4d7\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.899705 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-server-conf\") pod \"234155a2-9751-459f-b259-6bb32eb4b4d7\" (UID: \"234155a2-9751-459f-b259-6bb32eb4b4d7\") " Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.900788 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "234155a2-9751-459f-b259-6bb32eb4b4d7" (UID: "234155a2-9751-459f-b259-6bb32eb4b4d7"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.902369 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "234155a2-9751-459f-b259-6bb32eb4b4d7" (UID: "234155a2-9751-459f-b259-6bb32eb4b4d7"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.902913 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "234155a2-9751-459f-b259-6bb32eb4b4d7" (UID: "234155a2-9751-459f-b259-6bb32eb4b4d7"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.906491 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/234155a2-9751-459f-b259-6bb32eb4b4d7-pod-info" (OuterVolumeSpecName: "pod-info") pod "234155a2-9751-459f-b259-6bb32eb4b4d7" (UID: "234155a2-9751-459f-b259-6bb32eb4b4d7"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.908480 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "234155a2-9751-459f-b259-6bb32eb4b4d7" (UID: "234155a2-9751-459f-b259-6bb32eb4b4d7"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.910094 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "234155a2-9751-459f-b259-6bb32eb4b4d7" (UID: "234155a2-9751-459f-b259-6bb32eb4b4d7"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.910690 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-kube-api-access-44b7l" (OuterVolumeSpecName: "kube-api-access-44b7l") pod "234155a2-9751-459f-b259-6bb32eb4b4d7" (UID: "234155a2-9751-459f-b259-6bb32eb4b4d7"). InnerVolumeSpecName "kube-api-access-44b7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.911977 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/234155a2-9751-459f-b259-6bb32eb4b4d7-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "234155a2-9751-459f-b259-6bb32eb4b4d7" (UID: "234155a2-9751-459f-b259-6bb32eb4b4d7"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.934060 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-config-data" (OuterVolumeSpecName: "config-data") pod "234155a2-9751-459f-b259-6bb32eb4b4d7" (UID: "234155a2-9751-459f-b259-6bb32eb4b4d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:40.965575 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-server-conf" (OuterVolumeSpecName: "server-conf") pod "234155a2-9751-459f-b259-6bb32eb4b4d7" (UID: "234155a2-9751-459f-b259-6bb32eb4b4d7"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.002832 4745 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.002853 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.002863 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44b7l\" (UniqueName: \"kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-kube-api-access-44b7l\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.002874 4745 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/234155a2-9751-459f-b259-6bb32eb4b4d7-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.002900 4745 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.002909 4745 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.002918 4745 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-server-conf\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.002926 4745 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/234155a2-9751-459f-b259-6bb32eb4b4d7-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.002936 4745 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.002944 4745 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/234155a2-9751-459f-b259-6bb32eb4b4d7-pod-info\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.024413 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "234155a2-9751-459f-b259-6bb32eb4b4d7" (UID: "234155a2-9751-459f-b259-6bb32eb4b4d7"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.026686 4745 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.104824 4745 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.104853 4745 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/234155a2-9751-459f-b259-6bb32eb4b4d7-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.339125 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-c6k75"] Oct 10 13:37:41 crc kubenswrapper[4745]: E1010 13:37:41.339543 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="234155a2-9751-459f-b259-6bb32eb4b4d7" containerName="setup-container" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.339555 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="234155a2-9751-459f-b259-6bb32eb4b4d7" containerName="setup-container" Oct 10 13:37:41 crc kubenswrapper[4745]: E1010 13:37:41.339575 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="234155a2-9751-459f-b259-6bb32eb4b4d7" containerName="rabbitmq" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.339582 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="234155a2-9751-459f-b259-6bb32eb4b4d7" containerName="rabbitmq" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.339808 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="234155a2-9751-459f-b259-6bb32eb4b4d7" containerName="rabbitmq" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.340761 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.342690 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.358726 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-c6k75"] Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.410112 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4726\" (UniqueName: \"kubernetes.io/projected/b94a7e9d-555f-4fa0-abdb-21579a932bf9-kube-api-access-n4726\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.410351 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.410549 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-config\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.410596 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-dns-svc\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.410827 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.411030 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.411066 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.512702 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-config\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.512794 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-dns-svc\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.512872 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.512953 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.512979 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.513015 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4726\" (UniqueName: \"kubernetes.io/projected/b94a7e9d-555f-4fa0-abdb-21579a932bf9-kube-api-access-n4726\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.513039 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.513639 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-config\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.513962 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-dns-svc\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.514069 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.514095 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.514171 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.514189 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.551869 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4726\" (UniqueName: \"kubernetes.io/projected/b94a7e9d-555f-4fa0-abdb-21579a932bf9-kube-api-access-n4726\") pod \"dnsmasq-dns-67b789f86c-c6k75\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.665946 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.723636 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.723634 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"234155a2-9751-459f-b259-6bb32eb4b4d7","Type":"ContainerDied","Data":"3b8c5a01627d4585f00f719d1bd453a1eac0f2e8bae0eeabd5c4e20f8ff2eaae"} Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.723782 4745 scope.go:117] "RemoveContainer" containerID="c7ea1a57fde8811da41997a7bcd0d2d8f76592e019f75dd145aad49f9c8962d6" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.725969 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1","Type":"ContainerStarted","Data":"256424d2e1d92e4d5d103fc329cecfb8ac09413fb85d0ed65aa70db5111a7468"} Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.800667 4745 scope.go:117] "RemoveContainer" containerID="ba8c59cb21a1b92f873d2755789e9b4ca0999c1be1d4d2bdf990e767e13ffec7" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.829825 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.848836 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.862143 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.863958 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.866286 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.866718 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.867572 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.870022 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-hs5bv" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.870066 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.870177 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.874078 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 10 13:37:41 crc kubenswrapper[4745]: I1010 13:37:41.874192 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.019965 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/972f7a23-bb59-4b6c-8ee7-a51647bbc261-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.020020 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/972f7a23-bb59-4b6c-8ee7-a51647bbc261-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.020060 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/972f7a23-bb59-4b6c-8ee7-a51647bbc261-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.020088 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/972f7a23-bb59-4b6c-8ee7-a51647bbc261-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.020118 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/972f7a23-bb59-4b6c-8ee7-a51647bbc261-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.020198 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/972f7a23-bb59-4b6c-8ee7-a51647bbc261-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.020244 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.020318 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/972f7a23-bb59-4b6c-8ee7-a51647bbc261-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.020433 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/972f7a23-bb59-4b6c-8ee7-a51647bbc261-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.020556 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/972f7a23-bb59-4b6c-8ee7-a51647bbc261-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.020582 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8zml\" (UniqueName: \"kubernetes.io/projected/972f7a23-bb59-4b6c-8ee7-a51647bbc261-kube-api-access-v8zml\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.122894 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/972f7a23-bb59-4b6c-8ee7-a51647bbc261-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.122948 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/972f7a23-bb59-4b6c-8ee7-a51647bbc261-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.122985 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/972f7a23-bb59-4b6c-8ee7-a51647bbc261-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.123013 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/972f7a23-bb59-4b6c-8ee7-a51647bbc261-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.123041 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/972f7a23-bb59-4b6c-8ee7-a51647bbc261-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.123062 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/972f7a23-bb59-4b6c-8ee7-a51647bbc261-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.123085 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.123119 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/972f7a23-bb59-4b6c-8ee7-a51647bbc261-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.123248 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/972f7a23-bb59-4b6c-8ee7-a51647bbc261-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.123337 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/972f7a23-bb59-4b6c-8ee7-a51647bbc261-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.123416 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8zml\" (UniqueName: \"kubernetes.io/projected/972f7a23-bb59-4b6c-8ee7-a51647bbc261-kube-api-access-v8zml\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.125604 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/972f7a23-bb59-4b6c-8ee7-a51647bbc261-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.125699 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/972f7a23-bb59-4b6c-8ee7-a51647bbc261-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.126080 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/972f7a23-bb59-4b6c-8ee7-a51647bbc261-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.126276 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/972f7a23-bb59-4b6c-8ee7-a51647bbc261-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.126404 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.126889 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/972f7a23-bb59-4b6c-8ee7-a51647bbc261-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.128923 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/972f7a23-bb59-4b6c-8ee7-a51647bbc261-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.129784 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/972f7a23-bb59-4b6c-8ee7-a51647bbc261-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.130110 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/972f7a23-bb59-4b6c-8ee7-a51647bbc261-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.131092 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/972f7a23-bb59-4b6c-8ee7-a51647bbc261-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.142035 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8zml\" (UniqueName: \"kubernetes.io/projected/972f7a23-bb59-4b6c-8ee7-a51647bbc261-kube-api-access-v8zml\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.158264 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"972f7a23-bb59-4b6c-8ee7-a51647bbc261\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.171319 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-c6k75"] Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.190859 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.692390 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.741433 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"972f7a23-bb59-4b6c-8ee7-a51647bbc261","Type":"ContainerStarted","Data":"6fddc2684f9dfc63e461f7eb3b9a270d99e324b2dfb2bdf44622692940ab4dd6"} Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.743179 4745 generic.go:334] "Generic (PLEG): container finished" podID="b94a7e9d-555f-4fa0-abdb-21579a932bf9" containerID="4d2c9bf356d55cc6bf5332bac9360ac3d43f97fd0cef022974f4eca9f7fc1df6" exitCode=0 Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.743234 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-c6k75" event={"ID":"b94a7e9d-555f-4fa0-abdb-21579a932bf9","Type":"ContainerDied","Data":"4d2c9bf356d55cc6bf5332bac9360ac3d43f97fd0cef022974f4eca9f7fc1df6"} Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.743251 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-c6k75" event={"ID":"b94a7e9d-555f-4fa0-abdb-21579a932bf9","Type":"ContainerStarted","Data":"ca4a2349479920a31063895142f42c5d0de4ef88382147649cadff160853a82d"} Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.756417 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="234155a2-9751-459f-b259-6bb32eb4b4d7" path="/var/lib/kubelet/pods/234155a2-9751-459f-b259-6bb32eb4b4d7/volumes" Oct 10 13:37:42 crc kubenswrapper[4745]: I1010 13:37:42.757968 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1","Type":"ContainerStarted","Data":"0ee4949834bc04e17c217aa0859801c045421b19eb85a595f221ac8002de8c43"} Oct 10 13:37:43 crc kubenswrapper[4745]: I1010 13:37:43.762708 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-c6k75" event={"ID":"b94a7e9d-555f-4fa0-abdb-21579a932bf9","Type":"ContainerStarted","Data":"6b2ac44d6692dc9f6dcf65f005e63b8b1c33baea35cee6a8c63718a9870a9259"} Oct 10 13:37:43 crc kubenswrapper[4745]: I1010 13:37:43.802028 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b789f86c-c6k75" podStartSLOduration=2.8020033939999998 podStartE2EDuration="2.802003394s" podCreationTimestamp="2025-10-10 13:37:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:37:43.79636274 +0000 UTC m=+1177.694019533" watchObservedRunningTime="2025-10-10 13:37:43.802003394 +0000 UTC m=+1177.699660187" Oct 10 13:37:44 crc kubenswrapper[4745]: I1010 13:37:44.773157 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:45 crc kubenswrapper[4745]: I1010 13:37:45.789975 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"972f7a23-bb59-4b6c-8ee7-a51647bbc261","Type":"ContainerStarted","Data":"38f4070c5cced171c4516031ab0684287666134f6241de46c43fca8a1259b633"} Oct 10 13:37:51 crc kubenswrapper[4745]: I1010 13:37:51.666924 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:37:51 crc kubenswrapper[4745]: I1010 13:37:51.745322 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-mlmn9"] Oct 10 13:37:51 crc kubenswrapper[4745]: I1010 13:37:51.745889 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" podUID="cd189c07-04d2-4077-95bd-81e3504117cf" containerName="dnsmasq-dns" containerID="cri-o://393276f07936731c5060a1e0cfe82ff3b7e2cd94fd1d6faf36b376ff5cb4f413" gracePeriod=10 Oct 10 13:37:51 crc kubenswrapper[4745]: I1010 13:37:51.900201 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-4hmj7"] Oct 10 13:37:51 crc kubenswrapper[4745]: I1010 13:37:51.907479 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:51 crc kubenswrapper[4745]: I1010 13:37:51.919855 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-4hmj7"] Oct 10 13:37:51 crc kubenswrapper[4745]: I1010 13:37:51.933797 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:51 crc kubenswrapper[4745]: I1010 13:37:51.933896 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:51 crc kubenswrapper[4745]: I1010 13:37:51.933985 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:51 crc kubenswrapper[4745]: I1010 13:37:51.934017 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:51 crc kubenswrapper[4745]: I1010 13:37:51.934036 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:51 crc kubenswrapper[4745]: I1010 13:37:51.934061 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qtvz\" (UniqueName: \"kubernetes.io/projected/b5d53599-f2c8-4120-b767-72793898c000-kube-api-access-6qtvz\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:51 crc kubenswrapper[4745]: I1010 13:37:51.934215 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-config\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.036275 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.036404 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.036611 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.036649 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.036683 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qtvz\" (UniqueName: \"kubernetes.io/projected/b5d53599-f2c8-4120-b767-72793898c000-kube-api-access-6qtvz\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.036717 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-config\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.036792 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.037589 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.037595 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.037595 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.038081 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.038152 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-config\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.038379 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5d53599-f2c8-4120-b767-72793898c000-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.066758 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qtvz\" (UniqueName: \"kubernetes.io/projected/b5d53599-f2c8-4120-b767-72793898c000-kube-api-access-6qtvz\") pod \"dnsmasq-dns-cb6ffcf87-4hmj7\" (UID: \"b5d53599-f2c8-4120-b767-72793898c000\") " pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.241139 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.351057 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.443575 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m4z\" (UniqueName: \"kubernetes.io/projected/cd189c07-04d2-4077-95bd-81e3504117cf-kube-api-access-x2m4z\") pod \"cd189c07-04d2-4077-95bd-81e3504117cf\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.443698 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-ovsdbserver-nb\") pod \"cd189c07-04d2-4077-95bd-81e3504117cf\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.444324 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-dns-swift-storage-0\") pod \"cd189c07-04d2-4077-95bd-81e3504117cf\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.444357 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-config\") pod \"cd189c07-04d2-4077-95bd-81e3504117cf\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.444387 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-ovsdbserver-sb\") pod \"cd189c07-04d2-4077-95bd-81e3504117cf\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.444464 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-dns-svc\") pod \"cd189c07-04d2-4077-95bd-81e3504117cf\" (UID: \"cd189c07-04d2-4077-95bd-81e3504117cf\") " Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.450652 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd189c07-04d2-4077-95bd-81e3504117cf-kube-api-access-x2m4z" (OuterVolumeSpecName: "kube-api-access-x2m4z") pod "cd189c07-04d2-4077-95bd-81e3504117cf" (UID: "cd189c07-04d2-4077-95bd-81e3504117cf"). InnerVolumeSpecName "kube-api-access-x2m4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.507380 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-config" (OuterVolumeSpecName: "config") pod "cd189c07-04d2-4077-95bd-81e3504117cf" (UID: "cd189c07-04d2-4077-95bd-81e3504117cf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.508002 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cd189c07-04d2-4077-95bd-81e3504117cf" (UID: "cd189c07-04d2-4077-95bd-81e3504117cf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.511993 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cd189c07-04d2-4077-95bd-81e3504117cf" (UID: "cd189c07-04d2-4077-95bd-81e3504117cf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.512200 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cd189c07-04d2-4077-95bd-81e3504117cf" (UID: "cd189c07-04d2-4077-95bd-81e3504117cf"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.521597 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cd189c07-04d2-4077-95bd-81e3504117cf" (UID: "cd189c07-04d2-4077-95bd-81e3504117cf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.546977 4745 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.547007 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.547016 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.547040 4745 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.547049 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m4z\" (UniqueName: \"kubernetes.io/projected/cd189c07-04d2-4077-95bd-81e3504117cf-kube-api-access-x2m4z\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.547060 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd189c07-04d2-4077-95bd-81e3504117cf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.765484 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-4hmj7"] Oct 10 13:37:52 crc kubenswrapper[4745]: W1010 13:37:52.769411 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5d53599_f2c8_4120_b767_72793898c000.slice/crio-9bdb89047697f1e864651e54c28c4e907ab21def1c4b2ccae6e027b033f976a4 WatchSource:0}: Error finding container 9bdb89047697f1e864651e54c28c4e907ab21def1c4b2ccae6e027b033f976a4: Status 404 returned error can't find the container with id 9bdb89047697f1e864651e54c28c4e907ab21def1c4b2ccae6e027b033f976a4 Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.871294 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" event={"ID":"b5d53599-f2c8-4120-b767-72793898c000","Type":"ContainerStarted","Data":"9bdb89047697f1e864651e54c28c4e907ab21def1c4b2ccae6e027b033f976a4"} Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.873333 4745 generic.go:334] "Generic (PLEG): container finished" podID="cd189c07-04d2-4077-95bd-81e3504117cf" containerID="393276f07936731c5060a1e0cfe82ff3b7e2cd94fd1d6faf36b376ff5cb4f413" exitCode=0 Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.873420 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" event={"ID":"cd189c07-04d2-4077-95bd-81e3504117cf","Type":"ContainerDied","Data":"393276f07936731c5060a1e0cfe82ff3b7e2cd94fd1d6faf36b376ff5cb4f413"} Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.873494 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" event={"ID":"cd189c07-04d2-4077-95bd-81e3504117cf","Type":"ContainerDied","Data":"3693139174f559438f1673903e159ed79786e64b9118da704a0ced3d630b2ea3"} Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.873526 4745 scope.go:117] "RemoveContainer" containerID="393276f07936731c5060a1e0cfe82ff3b7e2cd94fd1d6faf36b376ff5cb4f413" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.873552 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-mlmn9" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.970703 4745 scope.go:117] "RemoveContainer" containerID="3b6565cb907867419092b17abbf534adb76c6512d3d32baba23b0ceb87041078" Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.974257 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-mlmn9"] Oct 10 13:37:52 crc kubenswrapper[4745]: I1010 13:37:52.984214 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-mlmn9"] Oct 10 13:37:53 crc kubenswrapper[4745]: I1010 13:37:53.006565 4745 scope.go:117] "RemoveContainer" containerID="393276f07936731c5060a1e0cfe82ff3b7e2cd94fd1d6faf36b376ff5cb4f413" Oct 10 13:37:53 crc kubenswrapper[4745]: E1010 13:37:53.007149 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"393276f07936731c5060a1e0cfe82ff3b7e2cd94fd1d6faf36b376ff5cb4f413\": container with ID starting with 393276f07936731c5060a1e0cfe82ff3b7e2cd94fd1d6faf36b376ff5cb4f413 not found: ID does not exist" containerID="393276f07936731c5060a1e0cfe82ff3b7e2cd94fd1d6faf36b376ff5cb4f413" Oct 10 13:37:53 crc kubenswrapper[4745]: I1010 13:37:53.007182 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"393276f07936731c5060a1e0cfe82ff3b7e2cd94fd1d6faf36b376ff5cb4f413"} err="failed to get container status \"393276f07936731c5060a1e0cfe82ff3b7e2cd94fd1d6faf36b376ff5cb4f413\": rpc error: code = NotFound desc = could not find container \"393276f07936731c5060a1e0cfe82ff3b7e2cd94fd1d6faf36b376ff5cb4f413\": container with ID starting with 393276f07936731c5060a1e0cfe82ff3b7e2cd94fd1d6faf36b376ff5cb4f413 not found: ID does not exist" Oct 10 13:37:53 crc kubenswrapper[4745]: I1010 13:37:53.007299 4745 scope.go:117] "RemoveContainer" containerID="3b6565cb907867419092b17abbf534adb76c6512d3d32baba23b0ceb87041078" Oct 10 13:37:53 crc kubenswrapper[4745]: E1010 13:37:53.009434 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b6565cb907867419092b17abbf534adb76c6512d3d32baba23b0ceb87041078\": container with ID starting with 3b6565cb907867419092b17abbf534adb76c6512d3d32baba23b0ceb87041078 not found: ID does not exist" containerID="3b6565cb907867419092b17abbf534adb76c6512d3d32baba23b0ceb87041078" Oct 10 13:37:53 crc kubenswrapper[4745]: I1010 13:37:53.009501 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b6565cb907867419092b17abbf534adb76c6512d3d32baba23b0ceb87041078"} err="failed to get container status \"3b6565cb907867419092b17abbf534adb76c6512d3d32baba23b0ceb87041078\": rpc error: code = NotFound desc = could not find container \"3b6565cb907867419092b17abbf534adb76c6512d3d32baba23b0ceb87041078\": container with ID starting with 3b6565cb907867419092b17abbf534adb76c6512d3d32baba23b0ceb87041078 not found: ID does not exist" Oct 10 13:37:53 crc kubenswrapper[4745]: I1010 13:37:53.887951 4745 generic.go:334] "Generic (PLEG): container finished" podID="b5d53599-f2c8-4120-b767-72793898c000" containerID="3ccc6da1a2aa1e8021a5f65c6188c4a45cdae01cf4940367868454c4024d3418" exitCode=0 Oct 10 13:37:53 crc kubenswrapper[4745]: I1010 13:37:53.888020 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" event={"ID":"b5d53599-f2c8-4120-b767-72793898c000","Type":"ContainerDied","Data":"3ccc6da1a2aa1e8021a5f65c6188c4a45cdae01cf4940367868454c4024d3418"} Oct 10 13:37:54 crc kubenswrapper[4745]: I1010 13:37:54.761598 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd189c07-04d2-4077-95bd-81e3504117cf" path="/var/lib/kubelet/pods/cd189c07-04d2-4077-95bd-81e3504117cf/volumes" Oct 10 13:37:54 crc kubenswrapper[4745]: I1010 13:37:54.904029 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" event={"ID":"b5d53599-f2c8-4120-b767-72793898c000","Type":"ContainerStarted","Data":"76beb7a379a668ad29a3d82d81a2d5afa8070f1a590bcddfaa40a18d2cb7fda9"} Oct 10 13:37:54 crc kubenswrapper[4745]: I1010 13:37:54.905591 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:37:54 crc kubenswrapper[4745]: I1010 13:37:54.934061 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" podStartSLOduration=3.934043514 podStartE2EDuration="3.934043514s" podCreationTimestamp="2025-10-10 13:37:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:37:54.929057017 +0000 UTC m=+1188.826713790" watchObservedRunningTime="2025-10-10 13:37:54.934043514 +0000 UTC m=+1188.831700277" Oct 10 13:38:02 crc kubenswrapper[4745]: I1010 13:38:02.243961 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb6ffcf87-4hmj7" Oct 10 13:38:02 crc kubenswrapper[4745]: I1010 13:38:02.347713 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-c6k75"] Oct 10 13:38:02 crc kubenswrapper[4745]: I1010 13:38:02.348090 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b789f86c-c6k75" podUID="b94a7e9d-555f-4fa0-abdb-21579a932bf9" containerName="dnsmasq-dns" containerID="cri-o://6b2ac44d6692dc9f6dcf65f005e63b8b1c33baea35cee6a8c63718a9870a9259" gracePeriod=10 Oct 10 13:38:02 crc kubenswrapper[4745]: I1010 13:38:02.803379 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:38:02 crc kubenswrapper[4745]: I1010 13:38:02.895608 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-dns-svc\") pod \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " Oct 10 13:38:02 crc kubenswrapper[4745]: I1010 13:38:02.895695 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-config\") pod \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " Oct 10 13:38:02 crc kubenswrapper[4745]: I1010 13:38:02.974359 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b94a7e9d-555f-4fa0-abdb-21579a932bf9" (UID: "b94a7e9d-555f-4fa0-abdb-21579a932bf9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:38:02 crc kubenswrapper[4745]: I1010 13:38:02.974929 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-config" (OuterVolumeSpecName: "config") pod "b94a7e9d-555f-4fa0-abdb-21579a932bf9" (UID: "b94a7e9d-555f-4fa0-abdb-21579a932bf9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:38:02 crc kubenswrapper[4745]: I1010 13:38:02.997452 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-ovsdbserver-nb\") pod \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " Oct 10 13:38:02 crc kubenswrapper[4745]: I1010 13:38:02.997512 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-openstack-edpm-ipam\") pod \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " Oct 10 13:38:02 crc kubenswrapper[4745]: I1010 13:38:02.997713 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-dns-swift-storage-0\") pod \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " Oct 10 13:38:02 crc kubenswrapper[4745]: I1010 13:38:02.997739 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-ovsdbserver-sb\") pod \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " Oct 10 13:38:02 crc kubenswrapper[4745]: I1010 13:38:02.997794 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4726\" (UniqueName: \"kubernetes.io/projected/b94a7e9d-555f-4fa0-abdb-21579a932bf9-kube-api-access-n4726\") pod \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\" (UID: \"b94a7e9d-555f-4fa0-abdb-21579a932bf9\") " Oct 10 13:38:02 crc kubenswrapper[4745]: I1010 13:38:02.998159 4745 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 13:38:02 crc kubenswrapper[4745]: I1010 13:38:02.998176 4745 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-config\") on node \"crc\" DevicePath \"\"" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.006616 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b94a7e9d-555f-4fa0-abdb-21579a932bf9-kube-api-access-n4726" (OuterVolumeSpecName: "kube-api-access-n4726") pod "b94a7e9d-555f-4fa0-abdb-21579a932bf9" (UID: "b94a7e9d-555f-4fa0-abdb-21579a932bf9"). InnerVolumeSpecName "kube-api-access-n4726". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.011252 4745 generic.go:334] "Generic (PLEG): container finished" podID="b94a7e9d-555f-4fa0-abdb-21579a932bf9" containerID="6b2ac44d6692dc9f6dcf65f005e63b8b1c33baea35cee6a8c63718a9870a9259" exitCode=0 Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.011295 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-c6k75" event={"ID":"b94a7e9d-555f-4fa0-abdb-21579a932bf9","Type":"ContainerDied","Data":"6b2ac44d6692dc9f6dcf65f005e63b8b1c33baea35cee6a8c63718a9870a9259"} Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.011314 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-c6k75" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.011338 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-c6k75" event={"ID":"b94a7e9d-555f-4fa0-abdb-21579a932bf9","Type":"ContainerDied","Data":"ca4a2349479920a31063895142f42c5d0de4ef88382147649cadff160853a82d"} Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.011361 4745 scope.go:117] "RemoveContainer" containerID="6b2ac44d6692dc9f6dcf65f005e63b8b1c33baea35cee6a8c63718a9870a9259" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.032296 4745 scope.go:117] "RemoveContainer" containerID="4d2c9bf356d55cc6bf5332bac9360ac3d43f97fd0cef022974f4eca9f7fc1df6" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.045122 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b94a7e9d-555f-4fa0-abdb-21579a932bf9" (UID: "b94a7e9d-555f-4fa0-abdb-21579a932bf9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.047309 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b94a7e9d-555f-4fa0-abdb-21579a932bf9" (UID: "b94a7e9d-555f-4fa0-abdb-21579a932bf9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.049344 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "b94a7e9d-555f-4fa0-abdb-21579a932bf9" (UID: "b94a7e9d-555f-4fa0-abdb-21579a932bf9"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.051556 4745 scope.go:117] "RemoveContainer" containerID="6b2ac44d6692dc9f6dcf65f005e63b8b1c33baea35cee6a8c63718a9870a9259" Oct 10 13:38:03 crc kubenswrapper[4745]: E1010 13:38:03.052039 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b2ac44d6692dc9f6dcf65f005e63b8b1c33baea35cee6a8c63718a9870a9259\": container with ID starting with 6b2ac44d6692dc9f6dcf65f005e63b8b1c33baea35cee6a8c63718a9870a9259 not found: ID does not exist" containerID="6b2ac44d6692dc9f6dcf65f005e63b8b1c33baea35cee6a8c63718a9870a9259" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.052076 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b2ac44d6692dc9f6dcf65f005e63b8b1c33baea35cee6a8c63718a9870a9259"} err="failed to get container status \"6b2ac44d6692dc9f6dcf65f005e63b8b1c33baea35cee6a8c63718a9870a9259\": rpc error: code = NotFound desc = could not find container \"6b2ac44d6692dc9f6dcf65f005e63b8b1c33baea35cee6a8c63718a9870a9259\": container with ID starting with 6b2ac44d6692dc9f6dcf65f005e63b8b1c33baea35cee6a8c63718a9870a9259 not found: ID does not exist" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.052129 4745 scope.go:117] "RemoveContainer" containerID="4d2c9bf356d55cc6bf5332bac9360ac3d43f97fd0cef022974f4eca9f7fc1df6" Oct 10 13:38:03 crc kubenswrapper[4745]: E1010 13:38:03.052608 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d2c9bf356d55cc6bf5332bac9360ac3d43f97fd0cef022974f4eca9f7fc1df6\": container with ID starting with 4d2c9bf356d55cc6bf5332bac9360ac3d43f97fd0cef022974f4eca9f7fc1df6 not found: ID does not exist" containerID="4d2c9bf356d55cc6bf5332bac9360ac3d43f97fd0cef022974f4eca9f7fc1df6" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.052659 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d2c9bf356d55cc6bf5332bac9360ac3d43f97fd0cef022974f4eca9f7fc1df6"} err="failed to get container status \"4d2c9bf356d55cc6bf5332bac9360ac3d43f97fd0cef022974f4eca9f7fc1df6\": rpc error: code = NotFound desc = could not find container \"4d2c9bf356d55cc6bf5332bac9360ac3d43f97fd0cef022974f4eca9f7fc1df6\": container with ID starting with 4d2c9bf356d55cc6bf5332bac9360ac3d43f97fd0cef022974f4eca9f7fc1df6 not found: ID does not exist" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.063238 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b94a7e9d-555f-4fa0-abdb-21579a932bf9" (UID: "b94a7e9d-555f-4fa0-abdb-21579a932bf9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.099604 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4726\" (UniqueName: \"kubernetes.io/projected/b94a7e9d-555f-4fa0-abdb-21579a932bf9-kube-api-access-n4726\") on node \"crc\" DevicePath \"\"" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.099634 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.099644 4745 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.099652 4745 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.099661 4745 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b94a7e9d-555f-4fa0-abdb-21579a932bf9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.344776 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-c6k75"] Oct 10 13:38:03 crc kubenswrapper[4745]: I1010 13:38:03.351509 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-c6k75"] Oct 10 13:38:04 crc kubenswrapper[4745]: I1010 13:38:04.758521 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b94a7e9d-555f-4fa0-abdb-21579a932bf9" path="/var/lib/kubelet/pods/b94a7e9d-555f-4fa0-abdb-21579a932bf9/volumes" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.142832 4745 generic.go:334] "Generic (PLEG): container finished" podID="3f7d0aba-ccad-4d20-9ab1-cf7d080571c1" containerID="0ee4949834bc04e17c217aa0859801c045421b19eb85a595f221ac8002de8c43" exitCode=0 Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.142899 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1","Type":"ContainerDied","Data":"0ee4949834bc04e17c217aa0859801c045421b19eb85a595f221ac8002de8c43"} Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.200850 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9"] Oct 10 13:38:16 crc kubenswrapper[4745]: E1010 13:38:16.201781 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b94a7e9d-555f-4fa0-abdb-21579a932bf9" containerName="dnsmasq-dns" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.201808 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="b94a7e9d-555f-4fa0-abdb-21579a932bf9" containerName="dnsmasq-dns" Oct 10 13:38:16 crc kubenswrapper[4745]: E1010 13:38:16.201837 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b94a7e9d-555f-4fa0-abdb-21579a932bf9" containerName="init" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.201847 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="b94a7e9d-555f-4fa0-abdb-21579a932bf9" containerName="init" Oct 10 13:38:16 crc kubenswrapper[4745]: E1010 13:38:16.201883 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd189c07-04d2-4077-95bd-81e3504117cf" containerName="dnsmasq-dns" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.201891 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd189c07-04d2-4077-95bd-81e3504117cf" containerName="dnsmasq-dns" Oct 10 13:38:16 crc kubenswrapper[4745]: E1010 13:38:16.201926 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd189c07-04d2-4077-95bd-81e3504117cf" containerName="init" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.201938 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd189c07-04d2-4077-95bd-81e3504117cf" containerName="init" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.202507 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="b94a7e9d-555f-4fa0-abdb-21579a932bf9" containerName="dnsmasq-dns" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.202539 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd189c07-04d2-4077-95bd-81e3504117cf" containerName="dnsmasq-dns" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.203636 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.211223 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.212149 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.212449 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.227409 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.249074 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9"] Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.354944 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9\" (UID: \"167c6043-6db1-4ead-8cf4-4040842b8d88\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.355150 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9\" (UID: \"167c6043-6db1-4ead-8cf4-4040842b8d88\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.355242 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9\" (UID: \"167c6043-6db1-4ead-8cf4-4040842b8d88\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.355313 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8wt7\" (UniqueName: \"kubernetes.io/projected/167c6043-6db1-4ead-8cf4-4040842b8d88-kube-api-access-f8wt7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9\" (UID: \"167c6043-6db1-4ead-8cf4-4040842b8d88\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.456940 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9\" (UID: \"167c6043-6db1-4ead-8cf4-4040842b8d88\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.457268 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9\" (UID: \"167c6043-6db1-4ead-8cf4-4040842b8d88\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.457320 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8wt7\" (UniqueName: \"kubernetes.io/projected/167c6043-6db1-4ead-8cf4-4040842b8d88-kube-api-access-f8wt7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9\" (UID: \"167c6043-6db1-4ead-8cf4-4040842b8d88\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.457361 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9\" (UID: \"167c6043-6db1-4ead-8cf4-4040842b8d88\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.464760 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9\" (UID: \"167c6043-6db1-4ead-8cf4-4040842b8d88\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.465724 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9\" (UID: \"167c6043-6db1-4ead-8cf4-4040842b8d88\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.466455 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9\" (UID: \"167c6043-6db1-4ead-8cf4-4040842b8d88\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.479884 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8wt7\" (UniqueName: \"kubernetes.io/projected/167c6043-6db1-4ead-8cf4-4040842b8d88-kube-api-access-f8wt7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9\" (UID: \"167c6043-6db1-4ead-8cf4-4040842b8d88\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" Oct 10 13:38:16 crc kubenswrapper[4745]: I1010 13:38:16.714171 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" Oct 10 13:38:17 crc kubenswrapper[4745]: I1010 13:38:17.158553 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3f7d0aba-ccad-4d20-9ab1-cf7d080571c1","Type":"ContainerStarted","Data":"964ecb9c4700ba972c279aea45987f6e739ae2aa720db5589a8614ad76f52c08"} Oct 10 13:38:17 crc kubenswrapper[4745]: I1010 13:38:17.159015 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 10 13:38:17 crc kubenswrapper[4745]: I1010 13:38:17.193750 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.193709993 podStartE2EDuration="37.193709993s" podCreationTimestamp="2025-10-10 13:37:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:38:17.183968785 +0000 UTC m=+1211.081625558" watchObservedRunningTime="2025-10-10 13:38:17.193709993 +0000 UTC m=+1211.091366756" Oct 10 13:38:17 crc kubenswrapper[4745]: I1010 13:38:17.285039 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9"] Oct 10 13:38:17 crc kubenswrapper[4745]: W1010 13:38:17.295252 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod167c6043_6db1_4ead_8cf4_4040842b8d88.slice/crio-d0b9d536037addc3f183a1eb6f1b489c67294f4bcef99a4b0ba64188e21e257e WatchSource:0}: Error finding container d0b9d536037addc3f183a1eb6f1b489c67294f4bcef99a4b0ba64188e21e257e: Status 404 returned error can't find the container with id d0b9d536037addc3f183a1eb6f1b489c67294f4bcef99a4b0ba64188e21e257e Oct 10 13:38:17 crc kubenswrapper[4745]: I1010 13:38:17.297914 4745 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 13:38:18 crc kubenswrapper[4745]: I1010 13:38:18.168501 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" event={"ID":"167c6043-6db1-4ead-8cf4-4040842b8d88","Type":"ContainerStarted","Data":"d0b9d536037addc3f183a1eb6f1b489c67294f4bcef99a4b0ba64188e21e257e"} Oct 10 13:38:19 crc kubenswrapper[4745]: I1010 13:38:19.177754 4745 generic.go:334] "Generic (PLEG): container finished" podID="972f7a23-bb59-4b6c-8ee7-a51647bbc261" containerID="38f4070c5cced171c4516031ab0684287666134f6241de46c43fca8a1259b633" exitCode=0 Oct 10 13:38:19 crc kubenswrapper[4745]: I1010 13:38:19.178019 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"972f7a23-bb59-4b6c-8ee7-a51647bbc261","Type":"ContainerDied","Data":"38f4070c5cced171c4516031ab0684287666134f6241de46c43fca8a1259b633"} Oct 10 13:38:20 crc kubenswrapper[4745]: I1010 13:38:20.190794 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"972f7a23-bb59-4b6c-8ee7-a51647bbc261","Type":"ContainerStarted","Data":"d28c070a7f40ab625b665a01798f48ce32a1f03ad8a5487ae01b8555b7fc790b"} Oct 10 13:38:20 crc kubenswrapper[4745]: I1010 13:38:20.191264 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:38:26 crc kubenswrapper[4745]: I1010 13:38:26.775044 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=45.775019316 podStartE2EDuration="45.775019316s" podCreationTimestamp="2025-10-10 13:37:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 13:38:20.219584855 +0000 UTC m=+1214.117241628" watchObservedRunningTime="2025-10-10 13:38:26.775019316 +0000 UTC m=+1220.672676099" Oct 10 13:38:28 crc kubenswrapper[4745]: I1010 13:38:28.260455 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" event={"ID":"167c6043-6db1-4ead-8cf4-4040842b8d88","Type":"ContainerStarted","Data":"41b16edb9606c0d3258d49a76216633a0f7d7f69fa391af107f86e06b02b4674"} Oct 10 13:38:28 crc kubenswrapper[4745]: I1010 13:38:28.304511 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" podStartSLOduration=2.3461328 podStartE2EDuration="12.304480422s" podCreationTimestamp="2025-10-10 13:38:16 +0000 UTC" firstStartedPulling="2025-10-10 13:38:17.297614772 +0000 UTC m=+1211.195271535" lastFinishedPulling="2025-10-10 13:38:27.255962394 +0000 UTC m=+1221.153619157" observedRunningTime="2025-10-10 13:38:28.276591834 +0000 UTC m=+1222.174248637" watchObservedRunningTime="2025-10-10 13:38:28.304480422 +0000 UTC m=+1222.202137225" Oct 10 13:38:30 crc kubenswrapper[4745]: I1010 13:38:30.396953 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 10 13:38:32 crc kubenswrapper[4745]: I1010 13:38:32.194987 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 10 13:38:39 crc kubenswrapper[4745]: I1010 13:38:39.401954 4745 generic.go:334] "Generic (PLEG): container finished" podID="167c6043-6db1-4ead-8cf4-4040842b8d88" containerID="41b16edb9606c0d3258d49a76216633a0f7d7f69fa391af107f86e06b02b4674" exitCode=0 Oct 10 13:38:39 crc kubenswrapper[4745]: I1010 13:38:39.402023 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" event={"ID":"167c6043-6db1-4ead-8cf4-4040842b8d88","Type":"ContainerDied","Data":"41b16edb9606c0d3258d49a76216633a0f7d7f69fa391af107f86e06b02b4674"} Oct 10 13:38:40 crc kubenswrapper[4745]: I1010 13:38:40.884080 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" Oct 10 13:38:40 crc kubenswrapper[4745]: I1010 13:38:40.992832 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-ssh-key\") pod \"167c6043-6db1-4ead-8cf4-4040842b8d88\" (UID: \"167c6043-6db1-4ead-8cf4-4040842b8d88\") " Oct 10 13:38:40 crc kubenswrapper[4745]: I1010 13:38:40.993025 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-inventory\") pod \"167c6043-6db1-4ead-8cf4-4040842b8d88\" (UID: \"167c6043-6db1-4ead-8cf4-4040842b8d88\") " Oct 10 13:38:40 crc kubenswrapper[4745]: I1010 13:38:40.993049 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-repo-setup-combined-ca-bundle\") pod \"167c6043-6db1-4ead-8cf4-4040842b8d88\" (UID: \"167c6043-6db1-4ead-8cf4-4040842b8d88\") " Oct 10 13:38:40 crc kubenswrapper[4745]: I1010 13:38:40.993086 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8wt7\" (UniqueName: \"kubernetes.io/projected/167c6043-6db1-4ead-8cf4-4040842b8d88-kube-api-access-f8wt7\") pod \"167c6043-6db1-4ead-8cf4-4040842b8d88\" (UID: \"167c6043-6db1-4ead-8cf4-4040842b8d88\") " Oct 10 13:38:40 crc kubenswrapper[4745]: I1010 13:38:40.999555 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/167c6043-6db1-4ead-8cf4-4040842b8d88-kube-api-access-f8wt7" (OuterVolumeSpecName: "kube-api-access-f8wt7") pod "167c6043-6db1-4ead-8cf4-4040842b8d88" (UID: "167c6043-6db1-4ead-8cf4-4040842b8d88"). InnerVolumeSpecName "kube-api-access-f8wt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.000697 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "167c6043-6db1-4ead-8cf4-4040842b8d88" (UID: "167c6043-6db1-4ead-8cf4-4040842b8d88"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.020923 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-inventory" (OuterVolumeSpecName: "inventory") pod "167c6043-6db1-4ead-8cf4-4040842b8d88" (UID: "167c6043-6db1-4ead-8cf4-4040842b8d88"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.032921 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "167c6043-6db1-4ead-8cf4-4040842b8d88" (UID: "167c6043-6db1-4ead-8cf4-4040842b8d88"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.095889 4745 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.095933 4745 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.095944 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8wt7\" (UniqueName: \"kubernetes.io/projected/167c6043-6db1-4ead-8cf4-4040842b8d88-kube-api-access-f8wt7\") on node \"crc\" DevicePath \"\"" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.095954 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/167c6043-6db1-4ead-8cf4-4040842b8d88-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.462881 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" event={"ID":"167c6043-6db1-4ead-8cf4-4040842b8d88","Type":"ContainerDied","Data":"d0b9d536037addc3f183a1eb6f1b489c67294f4bcef99a4b0ba64188e21e257e"} Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.462927 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0b9d536037addc3f183a1eb6f1b489c67294f4bcef99a4b0ba64188e21e257e" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.462984 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.515066 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw"] Oct 10 13:38:41 crc kubenswrapper[4745]: E1010 13:38:41.516992 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="167c6043-6db1-4ead-8cf4-4040842b8d88" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.517025 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="167c6043-6db1-4ead-8cf4-4040842b8d88" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.517268 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="167c6043-6db1-4ead-8cf4-4040842b8d88" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.518003 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.520890 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.521096 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.524297 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.524350 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.525382 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw"] Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.605593 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fvppw\" (UID: \"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.606005 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fvppw\" (UID: \"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.606089 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9nfw\" (UniqueName: \"kubernetes.io/projected/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-kube-api-access-k9nfw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fvppw\" (UID: \"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.709240 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fvppw\" (UID: \"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.709370 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fvppw\" (UID: \"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.709499 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9nfw\" (UniqueName: \"kubernetes.io/projected/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-kube-api-access-k9nfw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fvppw\" (UID: \"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.713195 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fvppw\" (UID: \"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.713519 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fvppw\" (UID: \"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.737468 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9nfw\" (UniqueName: \"kubernetes.io/projected/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-kube-api-access-k9nfw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fvppw\" (UID: \"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" Oct 10 13:38:41 crc kubenswrapper[4745]: I1010 13:38:41.855844 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" Oct 10 13:38:42 crc kubenswrapper[4745]: I1010 13:38:42.369765 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw"] Oct 10 13:38:42 crc kubenswrapper[4745]: I1010 13:38:42.474172 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" event={"ID":"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f","Type":"ContainerStarted","Data":"bc23156dcae83327b06adcb9e666167ddc316a99eaa1b4e6eeede740df2d15e6"} Oct 10 13:38:43 crc kubenswrapper[4745]: I1010 13:38:43.494119 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" event={"ID":"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f","Type":"ContainerStarted","Data":"646f26ca718fa98ef17c5a598b92d4b5de0d73e261b4c2c3cc06803e21ce012d"} Oct 10 13:38:43 crc kubenswrapper[4745]: I1010 13:38:43.528954 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" podStartSLOduration=2.104998701 podStartE2EDuration="2.528933929s" podCreationTimestamp="2025-10-10 13:38:41 +0000 UTC" firstStartedPulling="2025-10-10 13:38:42.37390137 +0000 UTC m=+1236.271558163" lastFinishedPulling="2025-10-10 13:38:42.797836598 +0000 UTC m=+1236.695493391" observedRunningTime="2025-10-10 13:38:43.517529749 +0000 UTC m=+1237.415186532" watchObservedRunningTime="2025-10-10 13:38:43.528933929 +0000 UTC m=+1237.426590702" Oct 10 13:38:45 crc kubenswrapper[4745]: I1010 13:38:45.524392 4745 generic.go:334] "Generic (PLEG): container finished" podID="4fda6419-bf48-4f0e-8ff8-da6b9c2da25f" containerID="646f26ca718fa98ef17c5a598b92d4b5de0d73e261b4c2c3cc06803e21ce012d" exitCode=0 Oct 10 13:38:45 crc kubenswrapper[4745]: I1010 13:38:45.524460 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" event={"ID":"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f","Type":"ContainerDied","Data":"646f26ca718fa98ef17c5a598b92d4b5de0d73e261b4c2c3cc06803e21ce012d"} Oct 10 13:38:46 crc kubenswrapper[4745]: I1010 13:38:46.187207 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:38:46 crc kubenswrapper[4745]: I1010 13:38:46.187332 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.008652 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.124102 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9nfw\" (UniqueName: \"kubernetes.io/projected/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-kube-api-access-k9nfw\") pod \"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f\" (UID: \"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f\") " Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.124208 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-ssh-key\") pod \"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f\" (UID: \"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f\") " Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.124276 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-inventory\") pod \"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f\" (UID: \"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f\") " Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.133371 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-kube-api-access-k9nfw" (OuterVolumeSpecName: "kube-api-access-k9nfw") pod "4fda6419-bf48-4f0e-8ff8-da6b9c2da25f" (UID: "4fda6419-bf48-4f0e-8ff8-da6b9c2da25f"). InnerVolumeSpecName "kube-api-access-k9nfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.162134 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-inventory" (OuterVolumeSpecName: "inventory") pod "4fda6419-bf48-4f0e-8ff8-da6b9c2da25f" (UID: "4fda6419-bf48-4f0e-8ff8-da6b9c2da25f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.175465 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4fda6419-bf48-4f0e-8ff8-da6b9c2da25f" (UID: "4fda6419-bf48-4f0e-8ff8-da6b9c2da25f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.227329 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9nfw\" (UniqueName: \"kubernetes.io/projected/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-kube-api-access-k9nfw\") on node \"crc\" DevicePath \"\"" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.227435 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.227450 4745 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fda6419-bf48-4f0e-8ff8-da6b9c2da25f-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.550495 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" event={"ID":"4fda6419-bf48-4f0e-8ff8-da6b9c2da25f","Type":"ContainerDied","Data":"bc23156dcae83327b06adcb9e666167ddc316a99eaa1b4e6eeede740df2d15e6"} Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.550547 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc23156dcae83327b06adcb9e666167ddc316a99eaa1b4e6eeede740df2d15e6" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.550593 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fvppw" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.628320 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh"] Oct 10 13:38:47 crc kubenswrapper[4745]: E1010 13:38:47.629016 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fda6419-bf48-4f0e-8ff8-da6b9c2da25f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.629041 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fda6419-bf48-4f0e-8ff8-da6b9c2da25f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.629457 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fda6419-bf48-4f0e-8ff8-da6b9c2da25f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.630579 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.640581 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.640894 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.641062 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.641259 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.647450 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh"] Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.736204 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh\" (UID: \"7f174876-f891-4508-97e6-e31624990bc8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.736648 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjkbc\" (UniqueName: \"kubernetes.io/projected/7f174876-f891-4508-97e6-e31624990bc8-kube-api-access-vjkbc\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh\" (UID: \"7f174876-f891-4508-97e6-e31624990bc8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.736723 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh\" (UID: \"7f174876-f891-4508-97e6-e31624990bc8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.736770 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh\" (UID: \"7f174876-f891-4508-97e6-e31624990bc8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.839021 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjkbc\" (UniqueName: \"kubernetes.io/projected/7f174876-f891-4508-97e6-e31624990bc8-kube-api-access-vjkbc\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh\" (UID: \"7f174876-f891-4508-97e6-e31624990bc8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.839534 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh\" (UID: \"7f174876-f891-4508-97e6-e31624990bc8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.840122 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh\" (UID: \"7f174876-f891-4508-97e6-e31624990bc8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.840584 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh\" (UID: \"7f174876-f891-4508-97e6-e31624990bc8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.844360 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh\" (UID: \"7f174876-f891-4508-97e6-e31624990bc8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.845037 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh\" (UID: \"7f174876-f891-4508-97e6-e31624990bc8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.847122 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh\" (UID: \"7f174876-f891-4508-97e6-e31624990bc8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.861652 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjkbc\" (UniqueName: \"kubernetes.io/projected/7f174876-f891-4508-97e6-e31624990bc8-kube-api-access-vjkbc\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh\" (UID: \"7f174876-f891-4508-97e6-e31624990bc8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" Oct 10 13:38:47 crc kubenswrapper[4745]: I1010 13:38:47.972718 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" Oct 10 13:38:48 crc kubenswrapper[4745]: I1010 13:38:48.526989 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh"] Oct 10 13:38:48 crc kubenswrapper[4745]: I1010 13:38:48.563777 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" event={"ID":"7f174876-f891-4508-97e6-e31624990bc8","Type":"ContainerStarted","Data":"364438d15aea8b355cdb8f9f29b3113dcd52296aeb7d1c8490a4c75047497738"} Oct 10 13:38:49 crc kubenswrapper[4745]: I1010 13:38:49.574148 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" event={"ID":"7f174876-f891-4508-97e6-e31624990bc8","Type":"ContainerStarted","Data":"2eba98fc937cddf82e59c0c999595a0e2f575ced01fa1c0ee8bdc8c588a366fd"} Oct 10 13:38:49 crc kubenswrapper[4745]: I1010 13:38:49.592214 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" podStartSLOduration=2.036186625 podStartE2EDuration="2.592190444s" podCreationTimestamp="2025-10-10 13:38:47 +0000 UTC" firstStartedPulling="2025-10-10 13:38:48.535386727 +0000 UTC m=+1242.433043540" lastFinishedPulling="2025-10-10 13:38:49.091390596 +0000 UTC m=+1242.989047359" observedRunningTime="2025-10-10 13:38:49.589979578 +0000 UTC m=+1243.487636361" watchObservedRunningTime="2025-10-10 13:38:49.592190444 +0000 UTC m=+1243.489847207" Oct 10 13:39:12 crc kubenswrapper[4745]: I1010 13:39:12.121471 4745 scope.go:117] "RemoveContainer" containerID="65ffcca13477e662350045a787cb182472c801c2ae1b42427a3267840f703e16" Oct 10 13:39:16 crc kubenswrapper[4745]: I1010 13:39:16.187235 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:39:16 crc kubenswrapper[4745]: I1010 13:39:16.188137 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:39:46 crc kubenswrapper[4745]: I1010 13:39:46.186816 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:39:46 crc kubenswrapper[4745]: I1010 13:39:46.187473 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:39:46 crc kubenswrapper[4745]: I1010 13:39:46.187537 4745 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:39:46 crc kubenswrapper[4745]: I1010 13:39:46.188499 4745 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"38f61d2405362274ab0410a7304f61bdd266d1575d542c8d7a802125d8ce76c0"} pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 13:39:46 crc kubenswrapper[4745]: I1010 13:39:46.188596 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" containerID="cri-o://38f61d2405362274ab0410a7304f61bdd266d1575d542c8d7a802125d8ce76c0" gracePeriod=600 Oct 10 13:39:47 crc kubenswrapper[4745]: I1010 13:39:47.268848 4745 generic.go:334] "Generic (PLEG): container finished" podID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerID="38f61d2405362274ab0410a7304f61bdd266d1575d542c8d7a802125d8ce76c0" exitCode=0 Oct 10 13:39:47 crc kubenswrapper[4745]: I1010 13:39:47.268929 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerDied","Data":"38f61d2405362274ab0410a7304f61bdd266d1575d542c8d7a802125d8ce76c0"} Oct 10 13:39:47 crc kubenswrapper[4745]: I1010 13:39:47.269594 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998"} Oct 10 13:39:47 crc kubenswrapper[4745]: I1010 13:39:47.269638 4745 scope.go:117] "RemoveContainer" containerID="051be97974334031c215babb8b965fd2290d7b1b439e8dabb7659be3724a636f" Oct 10 13:40:12 crc kubenswrapper[4745]: I1010 13:40:12.209195 4745 scope.go:117] "RemoveContainer" containerID="463a4a20101fbec8c7faee352813af08d95ca20f65c8903271ec715a712c7656" Oct 10 13:40:12 crc kubenswrapper[4745]: I1010 13:40:12.255154 4745 scope.go:117] "RemoveContainer" containerID="9daf00e75c695335dcf8fcee3fbe2dc11a393a7a3845077d6f3cf5f5383fe27a" Oct 10 13:40:12 crc kubenswrapper[4745]: I1010 13:40:12.342584 4745 scope.go:117] "RemoveContainer" containerID="17e01a5a08b518ada7b6fe6cf0842fb1a9ddeb2c4d93ec40c82f5f5354762385" Oct 10 13:41:12 crc kubenswrapper[4745]: I1010 13:41:12.494263 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hfr94"] Oct 10 13:41:12 crc kubenswrapper[4745]: I1010 13:41:12.499566 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hfr94" Oct 10 13:41:12 crc kubenswrapper[4745]: I1010 13:41:12.506641 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hfr94"] Oct 10 13:41:12 crc kubenswrapper[4745]: I1010 13:41:12.534007 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hmhj\" (UniqueName: \"kubernetes.io/projected/8edcb33d-1446-4590-ab65-be759496d4db-kube-api-access-5hmhj\") pod \"redhat-marketplace-hfr94\" (UID: \"8edcb33d-1446-4590-ab65-be759496d4db\") " pod="openshift-marketplace/redhat-marketplace-hfr94" Oct 10 13:41:12 crc kubenswrapper[4745]: I1010 13:41:12.534197 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8edcb33d-1446-4590-ab65-be759496d4db-catalog-content\") pod \"redhat-marketplace-hfr94\" (UID: \"8edcb33d-1446-4590-ab65-be759496d4db\") " pod="openshift-marketplace/redhat-marketplace-hfr94" Oct 10 13:41:12 crc kubenswrapper[4745]: I1010 13:41:12.534311 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8edcb33d-1446-4590-ab65-be759496d4db-utilities\") pod \"redhat-marketplace-hfr94\" (UID: \"8edcb33d-1446-4590-ab65-be759496d4db\") " pod="openshift-marketplace/redhat-marketplace-hfr94" Oct 10 13:41:12 crc kubenswrapper[4745]: I1010 13:41:12.635932 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hmhj\" (UniqueName: \"kubernetes.io/projected/8edcb33d-1446-4590-ab65-be759496d4db-kube-api-access-5hmhj\") pod \"redhat-marketplace-hfr94\" (UID: \"8edcb33d-1446-4590-ab65-be759496d4db\") " pod="openshift-marketplace/redhat-marketplace-hfr94" Oct 10 13:41:12 crc kubenswrapper[4745]: I1010 13:41:12.635985 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8edcb33d-1446-4590-ab65-be759496d4db-catalog-content\") pod \"redhat-marketplace-hfr94\" (UID: \"8edcb33d-1446-4590-ab65-be759496d4db\") " pod="openshift-marketplace/redhat-marketplace-hfr94" Oct 10 13:41:12 crc kubenswrapper[4745]: I1010 13:41:12.636039 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8edcb33d-1446-4590-ab65-be759496d4db-utilities\") pod \"redhat-marketplace-hfr94\" (UID: \"8edcb33d-1446-4590-ab65-be759496d4db\") " pod="openshift-marketplace/redhat-marketplace-hfr94" Oct 10 13:41:12 crc kubenswrapper[4745]: I1010 13:41:12.636562 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8edcb33d-1446-4590-ab65-be759496d4db-catalog-content\") pod \"redhat-marketplace-hfr94\" (UID: \"8edcb33d-1446-4590-ab65-be759496d4db\") " pod="openshift-marketplace/redhat-marketplace-hfr94" Oct 10 13:41:12 crc kubenswrapper[4745]: I1010 13:41:12.636583 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8edcb33d-1446-4590-ab65-be759496d4db-utilities\") pod \"redhat-marketplace-hfr94\" (UID: \"8edcb33d-1446-4590-ab65-be759496d4db\") " pod="openshift-marketplace/redhat-marketplace-hfr94" Oct 10 13:41:12 crc kubenswrapper[4745]: I1010 13:41:12.661413 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hmhj\" (UniqueName: \"kubernetes.io/projected/8edcb33d-1446-4590-ab65-be759496d4db-kube-api-access-5hmhj\") pod \"redhat-marketplace-hfr94\" (UID: \"8edcb33d-1446-4590-ab65-be759496d4db\") " pod="openshift-marketplace/redhat-marketplace-hfr94" Oct 10 13:41:12 crc kubenswrapper[4745]: I1010 13:41:12.870250 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hfr94" Oct 10 13:41:13 crc kubenswrapper[4745]: I1010 13:41:13.352560 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hfr94"] Oct 10 13:41:13 crc kubenswrapper[4745]: W1010 13:41:13.356863 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8edcb33d_1446_4590_ab65_be759496d4db.slice/crio-f0cb297930a9bb08077a91e5766246438be649e77fb2d0ba8f94b88ec484d417 WatchSource:0}: Error finding container f0cb297930a9bb08077a91e5766246438be649e77fb2d0ba8f94b88ec484d417: Status 404 returned error can't find the container with id f0cb297930a9bb08077a91e5766246438be649e77fb2d0ba8f94b88ec484d417 Oct 10 13:41:14 crc kubenswrapper[4745]: I1010 13:41:14.237926 4745 generic.go:334] "Generic (PLEG): container finished" podID="8edcb33d-1446-4590-ab65-be759496d4db" containerID="b7811b919aecb9acf3abdfd09bbcaf7b06888dc0478c375426c0fcd002e14c52" exitCode=0 Oct 10 13:41:14 crc kubenswrapper[4745]: I1010 13:41:14.238035 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hfr94" event={"ID":"8edcb33d-1446-4590-ab65-be759496d4db","Type":"ContainerDied","Data":"b7811b919aecb9acf3abdfd09bbcaf7b06888dc0478c375426c0fcd002e14c52"} Oct 10 13:41:14 crc kubenswrapper[4745]: I1010 13:41:14.238290 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hfr94" event={"ID":"8edcb33d-1446-4590-ab65-be759496d4db","Type":"ContainerStarted","Data":"f0cb297930a9bb08077a91e5766246438be649e77fb2d0ba8f94b88ec484d417"} Oct 10 13:41:16 crc kubenswrapper[4745]: I1010 13:41:16.261466 4745 generic.go:334] "Generic (PLEG): container finished" podID="8edcb33d-1446-4590-ab65-be759496d4db" containerID="c3fb8314c8c0cd8ca2dde2358a4d651e54225ba6e30dcfca7984facf51842823" exitCode=0 Oct 10 13:41:16 crc kubenswrapper[4745]: I1010 13:41:16.261510 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hfr94" event={"ID":"8edcb33d-1446-4590-ab65-be759496d4db","Type":"ContainerDied","Data":"c3fb8314c8c0cd8ca2dde2358a4d651e54225ba6e30dcfca7984facf51842823"} Oct 10 13:41:16 crc kubenswrapper[4745]: I1010 13:41:16.457005 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mv5rx"] Oct 10 13:41:16 crc kubenswrapper[4745]: I1010 13:41:16.460653 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mv5rx" Oct 10 13:41:16 crc kubenswrapper[4745]: I1010 13:41:16.469078 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mv5rx"] Oct 10 13:41:16 crc kubenswrapper[4745]: I1010 13:41:16.518214 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b644393c-820e-4b78-8ba4-e534e683bb72-utilities\") pod \"redhat-operators-mv5rx\" (UID: \"b644393c-820e-4b78-8ba4-e534e683bb72\") " pod="openshift-marketplace/redhat-operators-mv5rx" Oct 10 13:41:16 crc kubenswrapper[4745]: I1010 13:41:16.518303 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfms4\" (UniqueName: \"kubernetes.io/projected/b644393c-820e-4b78-8ba4-e534e683bb72-kube-api-access-dfms4\") pod \"redhat-operators-mv5rx\" (UID: \"b644393c-820e-4b78-8ba4-e534e683bb72\") " pod="openshift-marketplace/redhat-operators-mv5rx" Oct 10 13:41:16 crc kubenswrapper[4745]: I1010 13:41:16.518334 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b644393c-820e-4b78-8ba4-e534e683bb72-catalog-content\") pod \"redhat-operators-mv5rx\" (UID: \"b644393c-820e-4b78-8ba4-e534e683bb72\") " pod="openshift-marketplace/redhat-operators-mv5rx" Oct 10 13:41:16 crc kubenswrapper[4745]: I1010 13:41:16.620543 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b644393c-820e-4b78-8ba4-e534e683bb72-utilities\") pod \"redhat-operators-mv5rx\" (UID: \"b644393c-820e-4b78-8ba4-e534e683bb72\") " pod="openshift-marketplace/redhat-operators-mv5rx" Oct 10 13:41:16 crc kubenswrapper[4745]: I1010 13:41:16.620668 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfms4\" (UniqueName: \"kubernetes.io/projected/b644393c-820e-4b78-8ba4-e534e683bb72-kube-api-access-dfms4\") pod \"redhat-operators-mv5rx\" (UID: \"b644393c-820e-4b78-8ba4-e534e683bb72\") " pod="openshift-marketplace/redhat-operators-mv5rx" Oct 10 13:41:16 crc kubenswrapper[4745]: I1010 13:41:16.620748 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b644393c-820e-4b78-8ba4-e534e683bb72-catalog-content\") pod \"redhat-operators-mv5rx\" (UID: \"b644393c-820e-4b78-8ba4-e534e683bb72\") " pod="openshift-marketplace/redhat-operators-mv5rx" Oct 10 13:41:16 crc kubenswrapper[4745]: I1010 13:41:16.621222 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b644393c-820e-4b78-8ba4-e534e683bb72-utilities\") pod \"redhat-operators-mv5rx\" (UID: \"b644393c-820e-4b78-8ba4-e534e683bb72\") " pod="openshift-marketplace/redhat-operators-mv5rx" Oct 10 13:41:16 crc kubenswrapper[4745]: I1010 13:41:16.621390 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b644393c-820e-4b78-8ba4-e534e683bb72-catalog-content\") pod \"redhat-operators-mv5rx\" (UID: \"b644393c-820e-4b78-8ba4-e534e683bb72\") " pod="openshift-marketplace/redhat-operators-mv5rx" Oct 10 13:41:16 crc kubenswrapper[4745]: I1010 13:41:16.640177 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfms4\" (UniqueName: \"kubernetes.io/projected/b644393c-820e-4b78-8ba4-e534e683bb72-kube-api-access-dfms4\") pod \"redhat-operators-mv5rx\" (UID: \"b644393c-820e-4b78-8ba4-e534e683bb72\") " pod="openshift-marketplace/redhat-operators-mv5rx" Oct 10 13:41:16 crc kubenswrapper[4745]: I1010 13:41:16.799015 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mv5rx" Oct 10 13:41:17 crc kubenswrapper[4745]: I1010 13:41:17.278325 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hfr94" event={"ID":"8edcb33d-1446-4590-ab65-be759496d4db","Type":"ContainerStarted","Data":"d0292c4159856f20dbf554dd1b19177baa54e85e2b0917632e92ac2658f2ca9c"} Oct 10 13:41:17 crc kubenswrapper[4745]: I1010 13:41:17.297240 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mv5rx"] Oct 10 13:41:17 crc kubenswrapper[4745]: I1010 13:41:17.307825 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hfr94" podStartSLOduration=2.877880139 podStartE2EDuration="5.307800023s" podCreationTimestamp="2025-10-10 13:41:12 +0000 UTC" firstStartedPulling="2025-10-10 13:41:14.240352915 +0000 UTC m=+1388.138009678" lastFinishedPulling="2025-10-10 13:41:16.670272799 +0000 UTC m=+1390.567929562" observedRunningTime="2025-10-10 13:41:17.301248446 +0000 UTC m=+1391.198905219" watchObservedRunningTime="2025-10-10 13:41:17.307800023 +0000 UTC m=+1391.205456786" Oct 10 13:41:17 crc kubenswrapper[4745]: W1010 13:41:17.330230 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb644393c_820e_4b78_8ba4_e534e683bb72.slice/crio-5ccd33b08eaac2c5b26e1abf9d632c8a1ebc7754101a005b08a8c35cd78cf202 WatchSource:0}: Error finding container 5ccd33b08eaac2c5b26e1abf9d632c8a1ebc7754101a005b08a8c35cd78cf202: Status 404 returned error can't find the container with id 5ccd33b08eaac2c5b26e1abf9d632c8a1ebc7754101a005b08a8c35cd78cf202 Oct 10 13:41:18 crc kubenswrapper[4745]: I1010 13:41:18.291379 4745 generic.go:334] "Generic (PLEG): container finished" podID="b644393c-820e-4b78-8ba4-e534e683bb72" containerID="bfe976ed2398e8d77c13ac10ee3fcc98ddf8f542d4e55d4ae30f7cfd69f2c891" exitCode=0 Oct 10 13:41:18 crc kubenswrapper[4745]: I1010 13:41:18.291435 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv5rx" event={"ID":"b644393c-820e-4b78-8ba4-e534e683bb72","Type":"ContainerDied","Data":"bfe976ed2398e8d77c13ac10ee3fcc98ddf8f542d4e55d4ae30f7cfd69f2c891"} Oct 10 13:41:18 crc kubenswrapper[4745]: I1010 13:41:18.291872 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv5rx" event={"ID":"b644393c-820e-4b78-8ba4-e534e683bb72","Type":"ContainerStarted","Data":"5ccd33b08eaac2c5b26e1abf9d632c8a1ebc7754101a005b08a8c35cd78cf202"} Oct 10 13:41:20 crc kubenswrapper[4745]: I1010 13:41:20.321995 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv5rx" event={"ID":"b644393c-820e-4b78-8ba4-e534e683bb72","Type":"ContainerStarted","Data":"ba4e9025b51577836fdeb5c23adbd3166c6e1ead3298f980208b10e53b6632a9"} Oct 10 13:41:21 crc kubenswrapper[4745]: I1010 13:41:21.337713 4745 generic.go:334] "Generic (PLEG): container finished" podID="b644393c-820e-4b78-8ba4-e534e683bb72" containerID="ba4e9025b51577836fdeb5c23adbd3166c6e1ead3298f980208b10e53b6632a9" exitCode=0 Oct 10 13:41:21 crc kubenswrapper[4745]: I1010 13:41:21.337808 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv5rx" event={"ID":"b644393c-820e-4b78-8ba4-e534e683bb72","Type":"ContainerDied","Data":"ba4e9025b51577836fdeb5c23adbd3166c6e1ead3298f980208b10e53b6632a9"} Oct 10 13:41:22 crc kubenswrapper[4745]: I1010 13:41:22.353831 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv5rx" event={"ID":"b644393c-820e-4b78-8ba4-e534e683bb72","Type":"ContainerStarted","Data":"18b73c721c69552e30f335a1b0901a1fda61d9cad852c6920aaeef95b1453956"} Oct 10 13:41:22 crc kubenswrapper[4745]: I1010 13:41:22.373180 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mv5rx" podStartSLOduration=2.88160025 podStartE2EDuration="6.373149471s" podCreationTimestamp="2025-10-10 13:41:16 +0000 UTC" firstStartedPulling="2025-10-10 13:41:18.293079681 +0000 UTC m=+1392.190736444" lastFinishedPulling="2025-10-10 13:41:21.784628882 +0000 UTC m=+1395.682285665" observedRunningTime="2025-10-10 13:41:22.368898603 +0000 UTC m=+1396.266555436" watchObservedRunningTime="2025-10-10 13:41:22.373149471 +0000 UTC m=+1396.270806264" Oct 10 13:41:22 crc kubenswrapper[4745]: I1010 13:41:22.871418 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hfr94" Oct 10 13:41:22 crc kubenswrapper[4745]: I1010 13:41:22.871776 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hfr94" Oct 10 13:41:22 crc kubenswrapper[4745]: I1010 13:41:22.942634 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hfr94" Oct 10 13:41:23 crc kubenswrapper[4745]: I1010 13:41:23.417177 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hfr94" Oct 10 13:41:25 crc kubenswrapper[4745]: I1010 13:41:25.042339 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hfr94"] Oct 10 13:41:26 crc kubenswrapper[4745]: I1010 13:41:26.388328 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hfr94" podUID="8edcb33d-1446-4590-ab65-be759496d4db" containerName="registry-server" containerID="cri-o://d0292c4159856f20dbf554dd1b19177baa54e85e2b0917632e92ac2658f2ca9c" gracePeriod=2 Oct 10 13:41:26 crc kubenswrapper[4745]: I1010 13:41:26.799314 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mv5rx" Oct 10 13:41:26 crc kubenswrapper[4745]: I1010 13:41:26.799723 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mv5rx" Oct 10 13:41:26 crc kubenswrapper[4745]: I1010 13:41:26.872679 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hfr94" Oct 10 13:41:26 crc kubenswrapper[4745]: I1010 13:41:26.938469 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hmhj\" (UniqueName: \"kubernetes.io/projected/8edcb33d-1446-4590-ab65-be759496d4db-kube-api-access-5hmhj\") pod \"8edcb33d-1446-4590-ab65-be759496d4db\" (UID: \"8edcb33d-1446-4590-ab65-be759496d4db\") " Oct 10 13:41:26 crc kubenswrapper[4745]: I1010 13:41:26.938514 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8edcb33d-1446-4590-ab65-be759496d4db-catalog-content\") pod \"8edcb33d-1446-4590-ab65-be759496d4db\" (UID: \"8edcb33d-1446-4590-ab65-be759496d4db\") " Oct 10 13:41:26 crc kubenswrapper[4745]: I1010 13:41:26.938707 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8edcb33d-1446-4590-ab65-be759496d4db-utilities\") pod \"8edcb33d-1446-4590-ab65-be759496d4db\" (UID: \"8edcb33d-1446-4590-ab65-be759496d4db\") " Oct 10 13:41:26 crc kubenswrapper[4745]: I1010 13:41:26.939431 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8edcb33d-1446-4590-ab65-be759496d4db-utilities" (OuterVolumeSpecName: "utilities") pod "8edcb33d-1446-4590-ab65-be759496d4db" (UID: "8edcb33d-1446-4590-ab65-be759496d4db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:41:26 crc kubenswrapper[4745]: I1010 13:41:26.946143 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8edcb33d-1446-4590-ab65-be759496d4db-kube-api-access-5hmhj" (OuterVolumeSpecName: "kube-api-access-5hmhj") pod "8edcb33d-1446-4590-ab65-be759496d4db" (UID: "8edcb33d-1446-4590-ab65-be759496d4db"). InnerVolumeSpecName "kube-api-access-5hmhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:41:26 crc kubenswrapper[4745]: I1010 13:41:26.955602 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8edcb33d-1446-4590-ab65-be759496d4db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8edcb33d-1446-4590-ab65-be759496d4db" (UID: "8edcb33d-1446-4590-ab65-be759496d4db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.040519 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8edcb33d-1446-4590-ab65-be759496d4db-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.040548 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hmhj\" (UniqueName: \"kubernetes.io/projected/8edcb33d-1446-4590-ab65-be759496d4db-kube-api-access-5hmhj\") on node \"crc\" DevicePath \"\"" Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.040562 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8edcb33d-1446-4590-ab65-be759496d4db-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.398759 4745 generic.go:334] "Generic (PLEG): container finished" podID="8edcb33d-1446-4590-ab65-be759496d4db" containerID="d0292c4159856f20dbf554dd1b19177baa54e85e2b0917632e92ac2658f2ca9c" exitCode=0 Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.398802 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hfr94" event={"ID":"8edcb33d-1446-4590-ab65-be759496d4db","Type":"ContainerDied","Data":"d0292c4159856f20dbf554dd1b19177baa54e85e2b0917632e92ac2658f2ca9c"} Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.398809 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hfr94" Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.398827 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hfr94" event={"ID":"8edcb33d-1446-4590-ab65-be759496d4db","Type":"ContainerDied","Data":"f0cb297930a9bb08077a91e5766246438be649e77fb2d0ba8f94b88ec484d417"} Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.398845 4745 scope.go:117] "RemoveContainer" containerID="d0292c4159856f20dbf554dd1b19177baa54e85e2b0917632e92ac2658f2ca9c" Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.430611 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hfr94"] Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.432791 4745 scope.go:117] "RemoveContainer" containerID="c3fb8314c8c0cd8ca2dde2358a4d651e54225ba6e30dcfca7984facf51842823" Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.439598 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hfr94"] Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.459279 4745 scope.go:117] "RemoveContainer" containerID="b7811b919aecb9acf3abdfd09bbcaf7b06888dc0478c375426c0fcd002e14c52" Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.528600 4745 scope.go:117] "RemoveContainer" containerID="d0292c4159856f20dbf554dd1b19177baa54e85e2b0917632e92ac2658f2ca9c" Oct 10 13:41:27 crc kubenswrapper[4745]: E1010 13:41:27.529273 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0292c4159856f20dbf554dd1b19177baa54e85e2b0917632e92ac2658f2ca9c\": container with ID starting with d0292c4159856f20dbf554dd1b19177baa54e85e2b0917632e92ac2658f2ca9c not found: ID does not exist" containerID="d0292c4159856f20dbf554dd1b19177baa54e85e2b0917632e92ac2658f2ca9c" Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.529304 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0292c4159856f20dbf554dd1b19177baa54e85e2b0917632e92ac2658f2ca9c"} err="failed to get container status \"d0292c4159856f20dbf554dd1b19177baa54e85e2b0917632e92ac2658f2ca9c\": rpc error: code = NotFound desc = could not find container \"d0292c4159856f20dbf554dd1b19177baa54e85e2b0917632e92ac2658f2ca9c\": container with ID starting with d0292c4159856f20dbf554dd1b19177baa54e85e2b0917632e92ac2658f2ca9c not found: ID does not exist" Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.529415 4745 scope.go:117] "RemoveContainer" containerID="c3fb8314c8c0cd8ca2dde2358a4d651e54225ba6e30dcfca7984facf51842823" Oct 10 13:41:27 crc kubenswrapper[4745]: E1010 13:41:27.529724 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3fb8314c8c0cd8ca2dde2358a4d651e54225ba6e30dcfca7984facf51842823\": container with ID starting with c3fb8314c8c0cd8ca2dde2358a4d651e54225ba6e30dcfca7984facf51842823 not found: ID does not exist" containerID="c3fb8314c8c0cd8ca2dde2358a4d651e54225ba6e30dcfca7984facf51842823" Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.529792 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3fb8314c8c0cd8ca2dde2358a4d651e54225ba6e30dcfca7984facf51842823"} err="failed to get container status \"c3fb8314c8c0cd8ca2dde2358a4d651e54225ba6e30dcfca7984facf51842823\": rpc error: code = NotFound desc = could not find container \"c3fb8314c8c0cd8ca2dde2358a4d651e54225ba6e30dcfca7984facf51842823\": container with ID starting with c3fb8314c8c0cd8ca2dde2358a4d651e54225ba6e30dcfca7984facf51842823 not found: ID does not exist" Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.529822 4745 scope.go:117] "RemoveContainer" containerID="b7811b919aecb9acf3abdfd09bbcaf7b06888dc0478c375426c0fcd002e14c52" Oct 10 13:41:27 crc kubenswrapper[4745]: E1010 13:41:27.530455 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7811b919aecb9acf3abdfd09bbcaf7b06888dc0478c375426c0fcd002e14c52\": container with ID starting with b7811b919aecb9acf3abdfd09bbcaf7b06888dc0478c375426c0fcd002e14c52 not found: ID does not exist" containerID="b7811b919aecb9acf3abdfd09bbcaf7b06888dc0478c375426c0fcd002e14c52" Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.530496 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7811b919aecb9acf3abdfd09bbcaf7b06888dc0478c375426c0fcd002e14c52"} err="failed to get container status \"b7811b919aecb9acf3abdfd09bbcaf7b06888dc0478c375426c0fcd002e14c52\": rpc error: code = NotFound desc = could not find container \"b7811b919aecb9acf3abdfd09bbcaf7b06888dc0478c375426c0fcd002e14c52\": container with ID starting with b7811b919aecb9acf3abdfd09bbcaf7b06888dc0478c375426c0fcd002e14c52 not found: ID does not exist" Oct 10 13:41:27 crc kubenswrapper[4745]: I1010 13:41:27.857591 4745 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mv5rx" podUID="b644393c-820e-4b78-8ba4-e534e683bb72" containerName="registry-server" probeResult="failure" output=< Oct 10 13:41:27 crc kubenswrapper[4745]: timeout: failed to connect service ":50051" within 1s Oct 10 13:41:27 crc kubenswrapper[4745]: > Oct 10 13:41:28 crc kubenswrapper[4745]: I1010 13:41:28.763915 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8edcb33d-1446-4590-ab65-be759496d4db" path="/var/lib/kubelet/pods/8edcb33d-1446-4590-ab65-be759496d4db/volumes" Oct 10 13:41:31 crc kubenswrapper[4745]: I1010 13:41:31.652203 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bt6b7"] Oct 10 13:41:31 crc kubenswrapper[4745]: E1010 13:41:31.653569 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8edcb33d-1446-4590-ab65-be759496d4db" containerName="extract-utilities" Oct 10 13:41:31 crc kubenswrapper[4745]: I1010 13:41:31.653606 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="8edcb33d-1446-4590-ab65-be759496d4db" containerName="extract-utilities" Oct 10 13:41:31 crc kubenswrapper[4745]: E1010 13:41:31.653654 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8edcb33d-1446-4590-ab65-be759496d4db" containerName="registry-server" Oct 10 13:41:31 crc kubenswrapper[4745]: I1010 13:41:31.653672 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="8edcb33d-1446-4590-ab65-be759496d4db" containerName="registry-server" Oct 10 13:41:31 crc kubenswrapper[4745]: E1010 13:41:31.653764 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8edcb33d-1446-4590-ab65-be759496d4db" containerName="extract-content" Oct 10 13:41:31 crc kubenswrapper[4745]: I1010 13:41:31.653783 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="8edcb33d-1446-4590-ab65-be759496d4db" containerName="extract-content" Oct 10 13:41:31 crc kubenswrapper[4745]: I1010 13:41:31.654256 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="8edcb33d-1446-4590-ab65-be759496d4db" containerName="registry-server" Oct 10 13:41:31 crc kubenswrapper[4745]: I1010 13:41:31.657609 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bt6b7" Oct 10 13:41:31 crc kubenswrapper[4745]: I1010 13:41:31.673352 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bt6b7"] Oct 10 13:41:31 crc kubenswrapper[4745]: I1010 13:41:31.762114 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-catalog-content\") pod \"certified-operators-bt6b7\" (UID: \"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63\") " pod="openshift-marketplace/certified-operators-bt6b7" Oct 10 13:41:31 crc kubenswrapper[4745]: I1010 13:41:31.762307 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c528b\" (UniqueName: \"kubernetes.io/projected/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-kube-api-access-c528b\") pod \"certified-operators-bt6b7\" (UID: \"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63\") " pod="openshift-marketplace/certified-operators-bt6b7" Oct 10 13:41:31 crc kubenswrapper[4745]: I1010 13:41:31.762464 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-utilities\") pod \"certified-operators-bt6b7\" (UID: \"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63\") " pod="openshift-marketplace/certified-operators-bt6b7" Oct 10 13:41:31 crc kubenswrapper[4745]: I1010 13:41:31.864724 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-catalog-content\") pod \"certified-operators-bt6b7\" (UID: \"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63\") " pod="openshift-marketplace/certified-operators-bt6b7" Oct 10 13:41:31 crc kubenswrapper[4745]: I1010 13:41:31.864840 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c528b\" (UniqueName: \"kubernetes.io/projected/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-kube-api-access-c528b\") pod \"certified-operators-bt6b7\" (UID: \"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63\") " pod="openshift-marketplace/certified-operators-bt6b7" Oct 10 13:41:31 crc kubenswrapper[4745]: I1010 13:41:31.864926 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-utilities\") pod \"certified-operators-bt6b7\" (UID: \"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63\") " pod="openshift-marketplace/certified-operators-bt6b7" Oct 10 13:41:31 crc kubenswrapper[4745]: I1010 13:41:31.866187 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-catalog-content\") pod \"certified-operators-bt6b7\" (UID: \"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63\") " pod="openshift-marketplace/certified-operators-bt6b7" Oct 10 13:41:31 crc kubenswrapper[4745]: I1010 13:41:31.866241 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-utilities\") pod \"certified-operators-bt6b7\" (UID: \"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63\") " pod="openshift-marketplace/certified-operators-bt6b7" Oct 10 13:41:31 crc kubenswrapper[4745]: I1010 13:41:31.893874 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c528b\" (UniqueName: \"kubernetes.io/projected/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-kube-api-access-c528b\") pod \"certified-operators-bt6b7\" (UID: \"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63\") " pod="openshift-marketplace/certified-operators-bt6b7" Oct 10 13:41:31 crc kubenswrapper[4745]: I1010 13:41:31.991151 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bt6b7" Oct 10 13:41:32 crc kubenswrapper[4745]: I1010 13:41:32.354323 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bt6b7"] Oct 10 13:41:32 crc kubenswrapper[4745]: W1010 13:41:32.382928 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15034a37_ec2a_4ae3_b5f8_b61ef1cdef63.slice/crio-632824ce240f52827a782859f3f00c3eccc740bbb23674ea8cf4a8e69aa5a847 WatchSource:0}: Error finding container 632824ce240f52827a782859f3f00c3eccc740bbb23674ea8cf4a8e69aa5a847: Status 404 returned error can't find the container with id 632824ce240f52827a782859f3f00c3eccc740bbb23674ea8cf4a8e69aa5a847 Oct 10 13:41:32 crc kubenswrapper[4745]: I1010 13:41:32.460831 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt6b7" event={"ID":"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63","Type":"ContainerStarted","Data":"632824ce240f52827a782859f3f00c3eccc740bbb23674ea8cf4a8e69aa5a847"} Oct 10 13:41:33 crc kubenswrapper[4745]: I1010 13:41:33.475082 4745 generic.go:334] "Generic (PLEG): container finished" podID="15034a37-ec2a-4ae3-b5f8-b61ef1cdef63" containerID="ca34d6d34948da50aa7b9205c3d5fbae5158e0347c3cddc327b1d0a61b58dbbb" exitCode=0 Oct 10 13:41:33 crc kubenswrapper[4745]: I1010 13:41:33.475212 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt6b7" event={"ID":"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63","Type":"ContainerDied","Data":"ca34d6d34948da50aa7b9205c3d5fbae5158e0347c3cddc327b1d0a61b58dbbb"} Oct 10 13:41:35 crc kubenswrapper[4745]: I1010 13:41:35.510314 4745 generic.go:334] "Generic (PLEG): container finished" podID="15034a37-ec2a-4ae3-b5f8-b61ef1cdef63" containerID="86bd0f6deadbe65a5b7d03210ad1796549bcb81b6f465d609624936030c6c52d" exitCode=0 Oct 10 13:41:35 crc kubenswrapper[4745]: I1010 13:41:35.510422 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt6b7" event={"ID":"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63","Type":"ContainerDied","Data":"86bd0f6deadbe65a5b7d03210ad1796549bcb81b6f465d609624936030c6c52d"} Oct 10 13:41:36 crc kubenswrapper[4745]: I1010 13:41:36.523928 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt6b7" event={"ID":"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63","Type":"ContainerStarted","Data":"23b114a6ac05723c28d3ef8523daf59bed1bf4204bf4205a8d8ec3d359584e8c"} Oct 10 13:41:36 crc kubenswrapper[4745]: I1010 13:41:36.549425 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bt6b7" podStartSLOduration=3.075414218 podStartE2EDuration="5.549404741s" podCreationTimestamp="2025-10-10 13:41:31 +0000 UTC" firstStartedPulling="2025-10-10 13:41:33.479132831 +0000 UTC m=+1407.376789594" lastFinishedPulling="2025-10-10 13:41:35.953123354 +0000 UTC m=+1409.850780117" observedRunningTime="2025-10-10 13:41:36.543225704 +0000 UTC m=+1410.440882467" watchObservedRunningTime="2025-10-10 13:41:36.549404741 +0000 UTC m=+1410.447061504" Oct 10 13:41:36 crc kubenswrapper[4745]: I1010 13:41:36.861380 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mv5rx" Oct 10 13:41:36 crc kubenswrapper[4745]: I1010 13:41:36.927528 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mv5rx" Oct 10 13:41:38 crc kubenswrapper[4745]: I1010 13:41:38.001003 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mv5rx"] Oct 10 13:41:38 crc kubenswrapper[4745]: I1010 13:41:38.545511 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mv5rx" podUID="b644393c-820e-4b78-8ba4-e534e683bb72" containerName="registry-server" containerID="cri-o://18b73c721c69552e30f335a1b0901a1fda61d9cad852c6920aaeef95b1453956" gracePeriod=2 Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.060617 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mv5rx" Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.223587 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b644393c-820e-4b78-8ba4-e534e683bb72-utilities\") pod \"b644393c-820e-4b78-8ba4-e534e683bb72\" (UID: \"b644393c-820e-4b78-8ba4-e534e683bb72\") " Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.223692 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b644393c-820e-4b78-8ba4-e534e683bb72-catalog-content\") pod \"b644393c-820e-4b78-8ba4-e534e683bb72\" (UID: \"b644393c-820e-4b78-8ba4-e534e683bb72\") " Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.223939 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfms4\" (UniqueName: \"kubernetes.io/projected/b644393c-820e-4b78-8ba4-e534e683bb72-kube-api-access-dfms4\") pod \"b644393c-820e-4b78-8ba4-e534e683bb72\" (UID: \"b644393c-820e-4b78-8ba4-e534e683bb72\") " Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.225224 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b644393c-820e-4b78-8ba4-e534e683bb72-utilities" (OuterVolumeSpecName: "utilities") pod "b644393c-820e-4b78-8ba4-e534e683bb72" (UID: "b644393c-820e-4b78-8ba4-e534e683bb72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.234450 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b644393c-820e-4b78-8ba4-e534e683bb72-kube-api-access-dfms4" (OuterVolumeSpecName: "kube-api-access-dfms4") pod "b644393c-820e-4b78-8ba4-e534e683bb72" (UID: "b644393c-820e-4b78-8ba4-e534e683bb72"). InnerVolumeSpecName "kube-api-access-dfms4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.327307 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b644393c-820e-4b78-8ba4-e534e683bb72-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.327345 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfms4\" (UniqueName: \"kubernetes.io/projected/b644393c-820e-4b78-8ba4-e534e683bb72-kube-api-access-dfms4\") on node \"crc\" DevicePath \"\"" Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.349570 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b644393c-820e-4b78-8ba4-e534e683bb72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b644393c-820e-4b78-8ba4-e534e683bb72" (UID: "b644393c-820e-4b78-8ba4-e534e683bb72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.429070 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b644393c-820e-4b78-8ba4-e534e683bb72-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.562818 4745 generic.go:334] "Generic (PLEG): container finished" podID="b644393c-820e-4b78-8ba4-e534e683bb72" containerID="18b73c721c69552e30f335a1b0901a1fda61d9cad852c6920aaeef95b1453956" exitCode=0 Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.562899 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv5rx" event={"ID":"b644393c-820e-4b78-8ba4-e534e683bb72","Type":"ContainerDied","Data":"18b73c721c69552e30f335a1b0901a1fda61d9cad852c6920aaeef95b1453956"} Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.562941 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mv5rx" Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.562979 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv5rx" event={"ID":"b644393c-820e-4b78-8ba4-e534e683bb72","Type":"ContainerDied","Data":"5ccd33b08eaac2c5b26e1abf9d632c8a1ebc7754101a005b08a8c35cd78cf202"} Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.563022 4745 scope.go:117] "RemoveContainer" containerID="18b73c721c69552e30f335a1b0901a1fda61d9cad852c6920aaeef95b1453956" Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.604656 4745 scope.go:117] "RemoveContainer" containerID="ba4e9025b51577836fdeb5c23adbd3166c6e1ead3298f980208b10e53b6632a9" Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.612039 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mv5rx"] Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.624645 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mv5rx"] Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.632279 4745 scope.go:117] "RemoveContainer" containerID="bfe976ed2398e8d77c13ac10ee3fcc98ddf8f542d4e55d4ae30f7cfd69f2c891" Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.681497 4745 scope.go:117] "RemoveContainer" containerID="18b73c721c69552e30f335a1b0901a1fda61d9cad852c6920aaeef95b1453956" Oct 10 13:41:39 crc kubenswrapper[4745]: E1010 13:41:39.682408 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18b73c721c69552e30f335a1b0901a1fda61d9cad852c6920aaeef95b1453956\": container with ID starting with 18b73c721c69552e30f335a1b0901a1fda61d9cad852c6920aaeef95b1453956 not found: ID does not exist" containerID="18b73c721c69552e30f335a1b0901a1fda61d9cad852c6920aaeef95b1453956" Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.682445 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18b73c721c69552e30f335a1b0901a1fda61d9cad852c6920aaeef95b1453956"} err="failed to get container status \"18b73c721c69552e30f335a1b0901a1fda61d9cad852c6920aaeef95b1453956\": rpc error: code = NotFound desc = could not find container \"18b73c721c69552e30f335a1b0901a1fda61d9cad852c6920aaeef95b1453956\": container with ID starting with 18b73c721c69552e30f335a1b0901a1fda61d9cad852c6920aaeef95b1453956 not found: ID does not exist" Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.682474 4745 scope.go:117] "RemoveContainer" containerID="ba4e9025b51577836fdeb5c23adbd3166c6e1ead3298f980208b10e53b6632a9" Oct 10 13:41:39 crc kubenswrapper[4745]: E1010 13:41:39.682871 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba4e9025b51577836fdeb5c23adbd3166c6e1ead3298f980208b10e53b6632a9\": container with ID starting with ba4e9025b51577836fdeb5c23adbd3166c6e1ead3298f980208b10e53b6632a9 not found: ID does not exist" containerID="ba4e9025b51577836fdeb5c23adbd3166c6e1ead3298f980208b10e53b6632a9" Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.682899 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba4e9025b51577836fdeb5c23adbd3166c6e1ead3298f980208b10e53b6632a9"} err="failed to get container status \"ba4e9025b51577836fdeb5c23adbd3166c6e1ead3298f980208b10e53b6632a9\": rpc error: code = NotFound desc = could not find container \"ba4e9025b51577836fdeb5c23adbd3166c6e1ead3298f980208b10e53b6632a9\": container with ID starting with ba4e9025b51577836fdeb5c23adbd3166c6e1ead3298f980208b10e53b6632a9 not found: ID does not exist" Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.682915 4745 scope.go:117] "RemoveContainer" containerID="bfe976ed2398e8d77c13ac10ee3fcc98ddf8f542d4e55d4ae30f7cfd69f2c891" Oct 10 13:41:39 crc kubenswrapper[4745]: E1010 13:41:39.683441 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfe976ed2398e8d77c13ac10ee3fcc98ddf8f542d4e55d4ae30f7cfd69f2c891\": container with ID starting with bfe976ed2398e8d77c13ac10ee3fcc98ddf8f542d4e55d4ae30f7cfd69f2c891 not found: ID does not exist" containerID="bfe976ed2398e8d77c13ac10ee3fcc98ddf8f542d4e55d4ae30f7cfd69f2c891" Oct 10 13:41:39 crc kubenswrapper[4745]: I1010 13:41:39.683520 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfe976ed2398e8d77c13ac10ee3fcc98ddf8f542d4e55d4ae30f7cfd69f2c891"} err="failed to get container status \"bfe976ed2398e8d77c13ac10ee3fcc98ddf8f542d4e55d4ae30f7cfd69f2c891\": rpc error: code = NotFound desc = could not find container \"bfe976ed2398e8d77c13ac10ee3fcc98ddf8f542d4e55d4ae30f7cfd69f2c891\": container with ID starting with bfe976ed2398e8d77c13ac10ee3fcc98ddf8f542d4e55d4ae30f7cfd69f2c891 not found: ID does not exist" Oct 10 13:41:40 crc kubenswrapper[4745]: I1010 13:41:40.759866 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b644393c-820e-4b78-8ba4-e534e683bb72" path="/var/lib/kubelet/pods/b644393c-820e-4b78-8ba4-e534e683bb72/volumes" Oct 10 13:41:41 crc kubenswrapper[4745]: I1010 13:41:41.992334 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bt6b7" Oct 10 13:41:41 crc kubenswrapper[4745]: I1010 13:41:41.992479 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bt6b7" Oct 10 13:41:42 crc kubenswrapper[4745]: I1010 13:41:42.080231 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bt6b7" Oct 10 13:41:42 crc kubenswrapper[4745]: I1010 13:41:42.646186 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bt6b7" Oct 10 13:41:44 crc kubenswrapper[4745]: I1010 13:41:44.399457 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bt6b7"] Oct 10 13:41:44 crc kubenswrapper[4745]: I1010 13:41:44.610558 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bt6b7" podUID="15034a37-ec2a-4ae3-b5f8-b61ef1cdef63" containerName="registry-server" containerID="cri-o://23b114a6ac05723c28d3ef8523daf59bed1bf4204bf4205a8d8ec3d359584e8c" gracePeriod=2 Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.158858 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bt6b7" Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.277548 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-utilities\") pod \"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63\" (UID: \"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63\") " Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.277609 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-catalog-content\") pod \"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63\" (UID: \"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63\") " Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.277713 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c528b\" (UniqueName: \"kubernetes.io/projected/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-kube-api-access-c528b\") pod \"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63\" (UID: \"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63\") " Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.278611 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-utilities" (OuterVolumeSpecName: "utilities") pod "15034a37-ec2a-4ae3-b5f8-b61ef1cdef63" (UID: "15034a37-ec2a-4ae3-b5f8-b61ef1cdef63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.287093 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-kube-api-access-c528b" (OuterVolumeSpecName: "kube-api-access-c528b") pod "15034a37-ec2a-4ae3-b5f8-b61ef1cdef63" (UID: "15034a37-ec2a-4ae3-b5f8-b61ef1cdef63"). InnerVolumeSpecName "kube-api-access-c528b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.339345 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15034a37-ec2a-4ae3-b5f8-b61ef1cdef63" (UID: "15034a37-ec2a-4ae3-b5f8-b61ef1cdef63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.379517 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.379548 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.379559 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c528b\" (UniqueName: \"kubernetes.io/projected/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63-kube-api-access-c528b\") on node \"crc\" DevicePath \"\"" Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.628713 4745 generic.go:334] "Generic (PLEG): container finished" podID="15034a37-ec2a-4ae3-b5f8-b61ef1cdef63" containerID="23b114a6ac05723c28d3ef8523daf59bed1bf4204bf4205a8d8ec3d359584e8c" exitCode=0 Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.628777 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bt6b7" Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.628810 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt6b7" event={"ID":"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63","Type":"ContainerDied","Data":"23b114a6ac05723c28d3ef8523daf59bed1bf4204bf4205a8d8ec3d359584e8c"} Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.628905 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt6b7" event={"ID":"15034a37-ec2a-4ae3-b5f8-b61ef1cdef63","Type":"ContainerDied","Data":"632824ce240f52827a782859f3f00c3eccc740bbb23674ea8cf4a8e69aa5a847"} Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.628953 4745 scope.go:117] "RemoveContainer" containerID="23b114a6ac05723c28d3ef8523daf59bed1bf4204bf4205a8d8ec3d359584e8c" Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.680719 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bt6b7"] Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.683256 4745 scope.go:117] "RemoveContainer" containerID="86bd0f6deadbe65a5b7d03210ad1796549bcb81b6f465d609624936030c6c52d" Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.711673 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bt6b7"] Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.727953 4745 scope.go:117] "RemoveContainer" containerID="ca34d6d34948da50aa7b9205c3d5fbae5158e0347c3cddc327b1d0a61b58dbbb" Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.787408 4745 scope.go:117] "RemoveContainer" containerID="23b114a6ac05723c28d3ef8523daf59bed1bf4204bf4205a8d8ec3d359584e8c" Oct 10 13:41:45 crc kubenswrapper[4745]: E1010 13:41:45.788402 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23b114a6ac05723c28d3ef8523daf59bed1bf4204bf4205a8d8ec3d359584e8c\": container with ID starting with 23b114a6ac05723c28d3ef8523daf59bed1bf4204bf4205a8d8ec3d359584e8c not found: ID does not exist" containerID="23b114a6ac05723c28d3ef8523daf59bed1bf4204bf4205a8d8ec3d359584e8c" Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.788503 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23b114a6ac05723c28d3ef8523daf59bed1bf4204bf4205a8d8ec3d359584e8c"} err="failed to get container status \"23b114a6ac05723c28d3ef8523daf59bed1bf4204bf4205a8d8ec3d359584e8c\": rpc error: code = NotFound desc = could not find container \"23b114a6ac05723c28d3ef8523daf59bed1bf4204bf4205a8d8ec3d359584e8c\": container with ID starting with 23b114a6ac05723c28d3ef8523daf59bed1bf4204bf4205a8d8ec3d359584e8c not found: ID does not exist" Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.788574 4745 scope.go:117] "RemoveContainer" containerID="86bd0f6deadbe65a5b7d03210ad1796549bcb81b6f465d609624936030c6c52d" Oct 10 13:41:45 crc kubenswrapper[4745]: E1010 13:41:45.790556 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86bd0f6deadbe65a5b7d03210ad1796549bcb81b6f465d609624936030c6c52d\": container with ID starting with 86bd0f6deadbe65a5b7d03210ad1796549bcb81b6f465d609624936030c6c52d not found: ID does not exist" containerID="86bd0f6deadbe65a5b7d03210ad1796549bcb81b6f465d609624936030c6c52d" Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.790608 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86bd0f6deadbe65a5b7d03210ad1796549bcb81b6f465d609624936030c6c52d"} err="failed to get container status \"86bd0f6deadbe65a5b7d03210ad1796549bcb81b6f465d609624936030c6c52d\": rpc error: code = NotFound desc = could not find container \"86bd0f6deadbe65a5b7d03210ad1796549bcb81b6f465d609624936030c6c52d\": container with ID starting with 86bd0f6deadbe65a5b7d03210ad1796549bcb81b6f465d609624936030c6c52d not found: ID does not exist" Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.790646 4745 scope.go:117] "RemoveContainer" containerID="ca34d6d34948da50aa7b9205c3d5fbae5158e0347c3cddc327b1d0a61b58dbbb" Oct 10 13:41:45 crc kubenswrapper[4745]: E1010 13:41:45.791473 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca34d6d34948da50aa7b9205c3d5fbae5158e0347c3cddc327b1d0a61b58dbbb\": container with ID starting with ca34d6d34948da50aa7b9205c3d5fbae5158e0347c3cddc327b1d0a61b58dbbb not found: ID does not exist" containerID="ca34d6d34948da50aa7b9205c3d5fbae5158e0347c3cddc327b1d0a61b58dbbb" Oct 10 13:41:45 crc kubenswrapper[4745]: I1010 13:41:45.791515 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca34d6d34948da50aa7b9205c3d5fbae5158e0347c3cddc327b1d0a61b58dbbb"} err="failed to get container status \"ca34d6d34948da50aa7b9205c3d5fbae5158e0347c3cddc327b1d0a61b58dbbb\": rpc error: code = NotFound desc = could not find container \"ca34d6d34948da50aa7b9205c3d5fbae5158e0347c3cddc327b1d0a61b58dbbb\": container with ID starting with ca34d6d34948da50aa7b9205c3d5fbae5158e0347c3cddc327b1d0a61b58dbbb not found: ID does not exist" Oct 10 13:41:46 crc kubenswrapper[4745]: I1010 13:41:46.187265 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:41:46 crc kubenswrapper[4745]: I1010 13:41:46.187333 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:41:46 crc kubenswrapper[4745]: I1010 13:41:46.772511 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15034a37-ec2a-4ae3-b5f8-b61ef1cdef63" path="/var/lib/kubelet/pods/15034a37-ec2a-4ae3-b5f8-b61ef1cdef63/volumes" Oct 10 13:41:50 crc kubenswrapper[4745]: I1010 13:41:50.691967 4745 generic.go:334] "Generic (PLEG): container finished" podID="7f174876-f891-4508-97e6-e31624990bc8" containerID="2eba98fc937cddf82e59c0c999595a0e2f575ced01fa1c0ee8bdc8c588a366fd" exitCode=0 Oct 10 13:41:50 crc kubenswrapper[4745]: I1010 13:41:50.692026 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" event={"ID":"7f174876-f891-4508-97e6-e31624990bc8","Type":"ContainerDied","Data":"2eba98fc937cddf82e59c0c999595a0e2f575ced01fa1c0ee8bdc8c588a366fd"} Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.119365 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.242447 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-inventory\") pod \"7f174876-f891-4508-97e6-e31624990bc8\" (UID: \"7f174876-f891-4508-97e6-e31624990bc8\") " Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.242533 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-bootstrap-combined-ca-bundle\") pod \"7f174876-f891-4508-97e6-e31624990bc8\" (UID: \"7f174876-f891-4508-97e6-e31624990bc8\") " Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.242672 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-ssh-key\") pod \"7f174876-f891-4508-97e6-e31624990bc8\" (UID: \"7f174876-f891-4508-97e6-e31624990bc8\") " Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.242878 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjkbc\" (UniqueName: \"kubernetes.io/projected/7f174876-f891-4508-97e6-e31624990bc8-kube-api-access-vjkbc\") pod \"7f174876-f891-4508-97e6-e31624990bc8\" (UID: \"7f174876-f891-4508-97e6-e31624990bc8\") " Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.248814 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f174876-f891-4508-97e6-e31624990bc8-kube-api-access-vjkbc" (OuterVolumeSpecName: "kube-api-access-vjkbc") pod "7f174876-f891-4508-97e6-e31624990bc8" (UID: "7f174876-f891-4508-97e6-e31624990bc8"). InnerVolumeSpecName "kube-api-access-vjkbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.261195 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "7f174876-f891-4508-97e6-e31624990bc8" (UID: "7f174876-f891-4508-97e6-e31624990bc8"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.276270 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-inventory" (OuterVolumeSpecName: "inventory") pod "7f174876-f891-4508-97e6-e31624990bc8" (UID: "7f174876-f891-4508-97e6-e31624990bc8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.282101 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7f174876-f891-4508-97e6-e31624990bc8" (UID: "7f174876-f891-4508-97e6-e31624990bc8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.345823 4745 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.345864 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.345879 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjkbc\" (UniqueName: \"kubernetes.io/projected/7f174876-f891-4508-97e6-e31624990bc8-kube-api-access-vjkbc\") on node \"crc\" DevicePath \"\"" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.345893 4745 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f174876-f891-4508-97e6-e31624990bc8-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.713374 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" event={"ID":"7f174876-f891-4508-97e6-e31624990bc8","Type":"ContainerDied","Data":"364438d15aea8b355cdb8f9f29b3113dcd52296aeb7d1c8490a4c75047497738"} Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.713411 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="364438d15aea8b355cdb8f9f29b3113dcd52296aeb7d1c8490a4c75047497738" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.713446 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.810896 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8"] Oct 10 13:41:52 crc kubenswrapper[4745]: E1010 13:41:52.811251 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b644393c-820e-4b78-8ba4-e534e683bb72" containerName="extract-utilities" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.811269 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="b644393c-820e-4b78-8ba4-e534e683bb72" containerName="extract-utilities" Oct 10 13:41:52 crc kubenswrapper[4745]: E1010 13:41:52.811290 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b644393c-820e-4b78-8ba4-e534e683bb72" containerName="registry-server" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.811297 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="b644393c-820e-4b78-8ba4-e534e683bb72" containerName="registry-server" Oct 10 13:41:52 crc kubenswrapper[4745]: E1010 13:41:52.811309 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f174876-f891-4508-97e6-e31624990bc8" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.811315 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f174876-f891-4508-97e6-e31624990bc8" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 10 13:41:52 crc kubenswrapper[4745]: E1010 13:41:52.811340 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15034a37-ec2a-4ae3-b5f8-b61ef1cdef63" containerName="registry-server" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.811346 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="15034a37-ec2a-4ae3-b5f8-b61ef1cdef63" containerName="registry-server" Oct 10 13:41:52 crc kubenswrapper[4745]: E1010 13:41:52.811358 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15034a37-ec2a-4ae3-b5f8-b61ef1cdef63" containerName="extract-content" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.811364 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="15034a37-ec2a-4ae3-b5f8-b61ef1cdef63" containerName="extract-content" Oct 10 13:41:52 crc kubenswrapper[4745]: E1010 13:41:52.811379 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15034a37-ec2a-4ae3-b5f8-b61ef1cdef63" containerName="extract-utilities" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.811385 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="15034a37-ec2a-4ae3-b5f8-b61ef1cdef63" containerName="extract-utilities" Oct 10 13:41:52 crc kubenswrapper[4745]: E1010 13:41:52.811392 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b644393c-820e-4b78-8ba4-e534e683bb72" containerName="extract-content" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.811397 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="b644393c-820e-4b78-8ba4-e534e683bb72" containerName="extract-content" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.811590 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="15034a37-ec2a-4ae3-b5f8-b61ef1cdef63" containerName="registry-server" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.811609 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="b644393c-820e-4b78-8ba4-e534e683bb72" containerName="registry-server" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.811620 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f174876-f891-4508-97e6-e31624990bc8" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.812239 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.814662 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.816036 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.816207 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.816341 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.828976 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8"] Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.957197 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8\" (UID: \"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.957331 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcpbc\" (UniqueName: \"kubernetes.io/projected/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-kube-api-access-qcpbc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8\" (UID: \"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" Oct 10 13:41:52 crc kubenswrapper[4745]: I1010 13:41:52.957416 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8\" (UID: \"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" Oct 10 13:41:53 crc kubenswrapper[4745]: I1010 13:41:53.059100 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8\" (UID: \"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" Oct 10 13:41:53 crc kubenswrapper[4745]: I1010 13:41:53.059211 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8\" (UID: \"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" Oct 10 13:41:53 crc kubenswrapper[4745]: I1010 13:41:53.059324 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcpbc\" (UniqueName: \"kubernetes.io/projected/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-kube-api-access-qcpbc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8\" (UID: \"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" Oct 10 13:41:53 crc kubenswrapper[4745]: I1010 13:41:53.064624 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8\" (UID: \"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" Oct 10 13:41:53 crc kubenswrapper[4745]: I1010 13:41:53.066315 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8\" (UID: \"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" Oct 10 13:41:53 crc kubenswrapper[4745]: I1010 13:41:53.083903 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcpbc\" (UniqueName: \"kubernetes.io/projected/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-kube-api-access-qcpbc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8\" (UID: \"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" Oct 10 13:41:53 crc kubenswrapper[4745]: I1010 13:41:53.127575 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" Oct 10 13:41:53 crc kubenswrapper[4745]: I1010 13:41:53.643370 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8"] Oct 10 13:41:53 crc kubenswrapper[4745]: I1010 13:41:53.723119 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" event={"ID":"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18","Type":"ContainerStarted","Data":"61ac50aee2ce804a9036a05c70a44d26f5195f195197022378940008142e3578"} Oct 10 13:41:54 crc kubenswrapper[4745]: I1010 13:41:54.734038 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" event={"ID":"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18","Type":"ContainerStarted","Data":"655adce735902766a08fa9aa5dab7a8bca060cd3f8c615446e0b86bb56e26e41"} Oct 10 13:41:54 crc kubenswrapper[4745]: I1010 13:41:54.760661 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" podStartSLOduration=2.302809866 podStartE2EDuration="2.760640005s" podCreationTimestamp="2025-10-10 13:41:52 +0000 UTC" firstStartedPulling="2025-10-10 13:41:53.651789419 +0000 UTC m=+1427.549446182" lastFinishedPulling="2025-10-10 13:41:54.109619558 +0000 UTC m=+1428.007276321" observedRunningTime="2025-10-10 13:41:54.752214381 +0000 UTC m=+1428.649871184" watchObservedRunningTime="2025-10-10 13:41:54.760640005 +0000 UTC m=+1428.658296768" Oct 10 13:42:16 crc kubenswrapper[4745]: I1010 13:42:16.186961 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:42:16 crc kubenswrapper[4745]: I1010 13:42:16.187873 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:42:39 crc kubenswrapper[4745]: I1010 13:42:39.062059 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-dtjmj"] Oct 10 13:42:39 crc kubenswrapper[4745]: I1010 13:42:39.073248 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-dtjmj"] Oct 10 13:42:40 crc kubenswrapper[4745]: I1010 13:42:40.755330 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c7311f9-2a50-444c-b90d-c86b0d6f887b" path="/var/lib/kubelet/pods/8c7311f9-2a50-444c-b90d-c86b0d6f887b/volumes" Oct 10 13:42:41 crc kubenswrapper[4745]: I1010 13:42:41.061235 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-nb8v8"] Oct 10 13:42:41 crc kubenswrapper[4745]: I1010 13:42:41.073808 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-nb8v8"] Oct 10 13:42:42 crc kubenswrapper[4745]: I1010 13:42:42.030335 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-fxz7c"] Oct 10 13:42:42 crc kubenswrapper[4745]: I1010 13:42:42.043061 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-fxz7c"] Oct 10 13:42:42 crc kubenswrapper[4745]: I1010 13:42:42.755831 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34dd525b-9501-4685-b460-9564eed74e76" path="/var/lib/kubelet/pods/34dd525b-9501-4685-b460-9564eed74e76/volumes" Oct 10 13:42:42 crc kubenswrapper[4745]: I1010 13:42:42.756609 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="805bb2a1-bff0-4fda-be96-766d55898d4a" path="/var/lib/kubelet/pods/805bb2a1-bff0-4fda-be96-766d55898d4a/volumes" Oct 10 13:42:46 crc kubenswrapper[4745]: I1010 13:42:46.186677 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:42:46 crc kubenswrapper[4745]: I1010 13:42:46.187382 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:42:46 crc kubenswrapper[4745]: I1010 13:42:46.187430 4745 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:42:46 crc kubenswrapper[4745]: I1010 13:42:46.188119 4745 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998"} pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 13:42:46 crc kubenswrapper[4745]: I1010 13:42:46.188168 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" containerID="cri-o://59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" gracePeriod=600 Oct 10 13:42:46 crc kubenswrapper[4745]: E1010 13:42:46.344843 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:42:46 crc kubenswrapper[4745]: I1010 13:42:46.352324 4745 generic.go:334] "Generic (PLEG): container finished" podID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" exitCode=0 Oct 10 13:42:46 crc kubenswrapper[4745]: I1010 13:42:46.352414 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerDied","Data":"59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998"} Oct 10 13:42:46 crc kubenswrapper[4745]: I1010 13:42:46.352511 4745 scope.go:117] "RemoveContainer" containerID="38f61d2405362274ab0410a7304f61bdd266d1575d542c8d7a802125d8ce76c0" Oct 10 13:42:46 crc kubenswrapper[4745]: I1010 13:42:46.353802 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:42:46 crc kubenswrapper[4745]: E1010 13:42:46.354315 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:42:53 crc kubenswrapper[4745]: I1010 13:42:53.029489 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c34c-account-create-2hp4m"] Oct 10 13:42:53 crc kubenswrapper[4745]: I1010 13:42:53.037307 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-eef2-account-create-qmwvh"] Oct 10 13:42:53 crc kubenswrapper[4745]: I1010 13:42:53.045914 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-c34c-account-create-2hp4m"] Oct 10 13:42:53 crc kubenswrapper[4745]: I1010 13:42:53.052376 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-eef2-account-create-qmwvh"] Oct 10 13:42:54 crc kubenswrapper[4745]: I1010 13:42:54.761924 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e42a569-8a3d-410c-8fd8-98eeee59ebe9" path="/var/lib/kubelet/pods/7e42a569-8a3d-410c-8fd8-98eeee59ebe9/volumes" Oct 10 13:42:54 crc kubenswrapper[4745]: I1010 13:42:54.764523 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f136f642-8772-41b9-88d8-b28bb1111fef" path="/var/lib/kubelet/pods/f136f642-8772-41b9-88d8-b28bb1111fef/volumes" Oct 10 13:42:56 crc kubenswrapper[4745]: E1010 13:42:56.282247 4745 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Oct 10 13:42:57 crc kubenswrapper[4745]: I1010 13:42:57.035017 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-218e-account-create-l8rbx"] Oct 10 13:42:57 crc kubenswrapper[4745]: I1010 13:42:57.044841 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-218e-account-create-l8rbx"] Oct 10 13:42:57 crc kubenswrapper[4745]: I1010 13:42:57.744848 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:42:57 crc kubenswrapper[4745]: E1010 13:42:57.745556 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:42:58 crc kubenswrapper[4745]: I1010 13:42:58.323906 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cn7lb"] Oct 10 13:42:58 crc kubenswrapper[4745]: I1010 13:42:58.325915 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cn7lb" Oct 10 13:42:58 crc kubenswrapper[4745]: I1010 13:42:58.336933 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cn7lb"] Oct 10 13:42:58 crc kubenswrapper[4745]: I1010 13:42:58.419269 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz7ch\" (UniqueName: \"kubernetes.io/projected/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-kube-api-access-lz7ch\") pod \"community-operators-cn7lb\" (UID: \"dbb20043-1930-4c2d-b0d9-0cc7403cfc38\") " pod="openshift-marketplace/community-operators-cn7lb" Oct 10 13:42:58 crc kubenswrapper[4745]: I1010 13:42:58.419429 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-catalog-content\") pod \"community-operators-cn7lb\" (UID: \"dbb20043-1930-4c2d-b0d9-0cc7403cfc38\") " pod="openshift-marketplace/community-operators-cn7lb" Oct 10 13:42:58 crc kubenswrapper[4745]: I1010 13:42:58.419455 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-utilities\") pod \"community-operators-cn7lb\" (UID: \"dbb20043-1930-4c2d-b0d9-0cc7403cfc38\") " pod="openshift-marketplace/community-operators-cn7lb" Oct 10 13:42:58 crc kubenswrapper[4745]: I1010 13:42:58.521051 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-catalog-content\") pod \"community-operators-cn7lb\" (UID: \"dbb20043-1930-4c2d-b0d9-0cc7403cfc38\") " pod="openshift-marketplace/community-operators-cn7lb" Oct 10 13:42:58 crc kubenswrapper[4745]: I1010 13:42:58.521107 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-utilities\") pod \"community-operators-cn7lb\" (UID: \"dbb20043-1930-4c2d-b0d9-0cc7403cfc38\") " pod="openshift-marketplace/community-operators-cn7lb" Oct 10 13:42:58 crc kubenswrapper[4745]: I1010 13:42:58.521171 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz7ch\" (UniqueName: \"kubernetes.io/projected/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-kube-api-access-lz7ch\") pod \"community-operators-cn7lb\" (UID: \"dbb20043-1930-4c2d-b0d9-0cc7403cfc38\") " pod="openshift-marketplace/community-operators-cn7lb" Oct 10 13:42:58 crc kubenswrapper[4745]: I1010 13:42:58.521616 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-catalog-content\") pod \"community-operators-cn7lb\" (UID: \"dbb20043-1930-4c2d-b0d9-0cc7403cfc38\") " pod="openshift-marketplace/community-operators-cn7lb" Oct 10 13:42:58 crc kubenswrapper[4745]: I1010 13:42:58.521631 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-utilities\") pod \"community-operators-cn7lb\" (UID: \"dbb20043-1930-4c2d-b0d9-0cc7403cfc38\") " pod="openshift-marketplace/community-operators-cn7lb" Oct 10 13:42:58 crc kubenswrapper[4745]: I1010 13:42:58.541044 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz7ch\" (UniqueName: \"kubernetes.io/projected/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-kube-api-access-lz7ch\") pod \"community-operators-cn7lb\" (UID: \"dbb20043-1930-4c2d-b0d9-0cc7403cfc38\") " pod="openshift-marketplace/community-operators-cn7lb" Oct 10 13:42:58 crc kubenswrapper[4745]: I1010 13:42:58.672047 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cn7lb" Oct 10 13:42:58 crc kubenswrapper[4745]: I1010 13:42:58.761358 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="981cf93b-d455-4364-8d16-26b5d8e8ac56" path="/var/lib/kubelet/pods/981cf93b-d455-4364-8d16-26b5d8e8ac56/volumes" Oct 10 13:42:59 crc kubenswrapper[4745]: I1010 13:42:59.164337 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cn7lb"] Oct 10 13:42:59 crc kubenswrapper[4745]: I1010 13:42:59.487796 4745 generic.go:334] "Generic (PLEG): container finished" podID="dbb20043-1930-4c2d-b0d9-0cc7403cfc38" containerID="38e2f19a464a2f2f9d60ea157546f879ee6e670f10cadf94056461ca07855949" exitCode=0 Oct 10 13:42:59 crc kubenswrapper[4745]: I1010 13:42:59.487946 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cn7lb" event={"ID":"dbb20043-1930-4c2d-b0d9-0cc7403cfc38","Type":"ContainerDied","Data":"38e2f19a464a2f2f9d60ea157546f879ee6e670f10cadf94056461ca07855949"} Oct 10 13:42:59 crc kubenswrapper[4745]: I1010 13:42:59.488377 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cn7lb" event={"ID":"dbb20043-1930-4c2d-b0d9-0cc7403cfc38","Type":"ContainerStarted","Data":"1bc693b991e65b14e9ae27a0bbf5e1255eb778d5865c1c13cd851c4c06a1195b"} Oct 10 13:43:01 crc kubenswrapper[4745]: I1010 13:43:01.512060 4745 generic.go:334] "Generic (PLEG): container finished" podID="dbb20043-1930-4c2d-b0d9-0cc7403cfc38" containerID="e7b761490e03630d5fe7c43e723fd846f14d15fbdd0c4d3df811f5ceb05c797b" exitCode=0 Oct 10 13:43:01 crc kubenswrapper[4745]: I1010 13:43:01.512210 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cn7lb" event={"ID":"dbb20043-1930-4c2d-b0d9-0cc7403cfc38","Type":"ContainerDied","Data":"e7b761490e03630d5fe7c43e723fd846f14d15fbdd0c4d3df811f5ceb05c797b"} Oct 10 13:43:02 crc kubenswrapper[4745]: I1010 13:43:02.526155 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cn7lb" event={"ID":"dbb20043-1930-4c2d-b0d9-0cc7403cfc38","Type":"ContainerStarted","Data":"6d07059da43a8db86e095c2e01bf028704e57d9b780d43607ed2407c619b18c0"} Oct 10 13:43:02 crc kubenswrapper[4745]: I1010 13:43:02.559415 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cn7lb" podStartSLOduration=1.963664748 podStartE2EDuration="4.559386673s" podCreationTimestamp="2025-10-10 13:42:58 +0000 UTC" firstStartedPulling="2025-10-10 13:42:59.492934431 +0000 UTC m=+1493.390591194" lastFinishedPulling="2025-10-10 13:43:02.088656356 +0000 UTC m=+1495.986313119" observedRunningTime="2025-10-10 13:43:02.5557175 +0000 UTC m=+1496.453374263" watchObservedRunningTime="2025-10-10 13:43:02.559386673 +0000 UTC m=+1496.457043436" Oct 10 13:43:08 crc kubenswrapper[4745]: I1010 13:43:08.672846 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cn7lb" Oct 10 13:43:08 crc kubenswrapper[4745]: I1010 13:43:08.673398 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cn7lb" Oct 10 13:43:08 crc kubenswrapper[4745]: I1010 13:43:08.743745 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cn7lb" Oct 10 13:43:08 crc kubenswrapper[4745]: I1010 13:43:08.744980 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:43:08 crc kubenswrapper[4745]: E1010 13:43:08.745250 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:43:09 crc kubenswrapper[4745]: I1010 13:43:09.670020 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cn7lb" Oct 10 13:43:09 crc kubenswrapper[4745]: I1010 13:43:09.731570 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cn7lb"] Oct 10 13:43:11 crc kubenswrapper[4745]: I1010 13:43:11.636004 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cn7lb" podUID="dbb20043-1930-4c2d-b0d9-0cc7403cfc38" containerName="registry-server" containerID="cri-o://6d07059da43a8db86e095c2e01bf028704e57d9b780d43607ed2407c619b18c0" gracePeriod=2 Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.108685 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cn7lb" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.225319 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-catalog-content\") pod \"dbb20043-1930-4c2d-b0d9-0cc7403cfc38\" (UID: \"dbb20043-1930-4c2d-b0d9-0cc7403cfc38\") " Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.225493 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz7ch\" (UniqueName: \"kubernetes.io/projected/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-kube-api-access-lz7ch\") pod \"dbb20043-1930-4c2d-b0d9-0cc7403cfc38\" (UID: \"dbb20043-1930-4c2d-b0d9-0cc7403cfc38\") " Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.225658 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-utilities\") pod \"dbb20043-1930-4c2d-b0d9-0cc7403cfc38\" (UID: \"dbb20043-1930-4c2d-b0d9-0cc7403cfc38\") " Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.226618 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-utilities" (OuterVolumeSpecName: "utilities") pod "dbb20043-1930-4c2d-b0d9-0cc7403cfc38" (UID: "dbb20043-1930-4c2d-b0d9-0cc7403cfc38"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.234320 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-kube-api-access-lz7ch" (OuterVolumeSpecName: "kube-api-access-lz7ch") pod "dbb20043-1930-4c2d-b0d9-0cc7403cfc38" (UID: "dbb20043-1930-4c2d-b0d9-0cc7403cfc38"). InnerVolumeSpecName "kube-api-access-lz7ch". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.328047 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz7ch\" (UniqueName: \"kubernetes.io/projected/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-kube-api-access-lz7ch\") on node \"crc\" DevicePath \"\"" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.328554 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.534523 4745 scope.go:117] "RemoveContainer" containerID="2ad421cb34f2c4e036e33aa1660436c3b3bc8a66e4c156aa432e25fb99ab3257" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.628783 4745 scope.go:117] "RemoveContainer" containerID="c4820afa2c40ee2f1bab3f2caac896b75a46ac51298b9e561885f1d740592ddf" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.652157 4745 generic.go:334] "Generic (PLEG): container finished" podID="dbb20043-1930-4c2d-b0d9-0cc7403cfc38" containerID="6d07059da43a8db86e095c2e01bf028704e57d9b780d43607ed2407c619b18c0" exitCode=0 Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.652286 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cn7lb" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.652290 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cn7lb" event={"ID":"dbb20043-1930-4c2d-b0d9-0cc7403cfc38","Type":"ContainerDied","Data":"6d07059da43a8db86e095c2e01bf028704e57d9b780d43607ed2407c619b18c0"} Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.653103 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cn7lb" event={"ID":"dbb20043-1930-4c2d-b0d9-0cc7403cfc38","Type":"ContainerDied","Data":"1bc693b991e65b14e9ae27a0bbf5e1255eb778d5865c1c13cd851c4c06a1195b"} Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.653165 4745 scope.go:117] "RemoveContainer" containerID="6d07059da43a8db86e095c2e01bf028704e57d9b780d43607ed2407c619b18c0" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.736083 4745 scope.go:117] "RemoveContainer" containerID="c73d4a837753e4d6f37bf1af6f7a91fcfb0909353cef4a00bbde1222cb38bf22" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.747556 4745 scope.go:117] "RemoveContainer" containerID="e7b761490e03630d5fe7c43e723fd846f14d15fbdd0c4d3df811f5ceb05c797b" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.768223 4745 scope.go:117] "RemoveContainer" containerID="3a519a98f25ce3069a56ae575dbf6156951176efb68ad1b733d2dac9e80c2e8d" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.791516 4745 scope.go:117] "RemoveContainer" containerID="38e2f19a464a2f2f9d60ea157546f879ee6e670f10cadf94056461ca07855949" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.888763 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dbb20043-1930-4c2d-b0d9-0cc7403cfc38" (UID: "dbb20043-1930-4c2d-b0d9-0cc7403cfc38"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.898943 4745 scope.go:117] "RemoveContainer" containerID="87dffe63db1888612bcf95d6783386baf4376547f0da3816f104721cd81fb8de" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.926705 4745 scope.go:117] "RemoveContainer" containerID="6d07059da43a8db86e095c2e01bf028704e57d9b780d43607ed2407c619b18c0" Oct 10 13:43:12 crc kubenswrapper[4745]: E1010 13:43:12.927661 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d07059da43a8db86e095c2e01bf028704e57d9b780d43607ed2407c619b18c0\": container with ID starting with 6d07059da43a8db86e095c2e01bf028704e57d9b780d43607ed2407c619b18c0 not found: ID does not exist" containerID="6d07059da43a8db86e095c2e01bf028704e57d9b780d43607ed2407c619b18c0" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.927813 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d07059da43a8db86e095c2e01bf028704e57d9b780d43607ed2407c619b18c0"} err="failed to get container status \"6d07059da43a8db86e095c2e01bf028704e57d9b780d43607ed2407c619b18c0\": rpc error: code = NotFound desc = could not find container \"6d07059da43a8db86e095c2e01bf028704e57d9b780d43607ed2407c619b18c0\": container with ID starting with 6d07059da43a8db86e095c2e01bf028704e57d9b780d43607ed2407c619b18c0 not found: ID does not exist" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.927884 4745 scope.go:117] "RemoveContainer" containerID="e7b761490e03630d5fe7c43e723fd846f14d15fbdd0c4d3df811f5ceb05c797b" Oct 10 13:43:12 crc kubenswrapper[4745]: E1010 13:43:12.934946 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7b761490e03630d5fe7c43e723fd846f14d15fbdd0c4d3df811f5ceb05c797b\": container with ID starting with e7b761490e03630d5fe7c43e723fd846f14d15fbdd0c4d3df811f5ceb05c797b not found: ID does not exist" containerID="e7b761490e03630d5fe7c43e723fd846f14d15fbdd0c4d3df811f5ceb05c797b" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.935003 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7b761490e03630d5fe7c43e723fd846f14d15fbdd0c4d3df811f5ceb05c797b"} err="failed to get container status \"e7b761490e03630d5fe7c43e723fd846f14d15fbdd0c4d3df811f5ceb05c797b\": rpc error: code = NotFound desc = could not find container \"e7b761490e03630d5fe7c43e723fd846f14d15fbdd0c4d3df811f5ceb05c797b\": container with ID starting with e7b761490e03630d5fe7c43e723fd846f14d15fbdd0c4d3df811f5ceb05c797b not found: ID does not exist" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.935035 4745 scope.go:117] "RemoveContainer" containerID="38e2f19a464a2f2f9d60ea157546f879ee6e670f10cadf94056461ca07855949" Oct 10 13:43:12 crc kubenswrapper[4745]: E1010 13:43:12.935620 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38e2f19a464a2f2f9d60ea157546f879ee6e670f10cadf94056461ca07855949\": container with ID starting with 38e2f19a464a2f2f9d60ea157546f879ee6e670f10cadf94056461ca07855949 not found: ID does not exist" containerID="38e2f19a464a2f2f9d60ea157546f879ee6e670f10cadf94056461ca07855949" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.935650 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38e2f19a464a2f2f9d60ea157546f879ee6e670f10cadf94056461ca07855949"} err="failed to get container status \"38e2f19a464a2f2f9d60ea157546f879ee6e670f10cadf94056461ca07855949\": rpc error: code = NotFound desc = could not find container \"38e2f19a464a2f2f9d60ea157546f879ee6e670f10cadf94056461ca07855949\": container with ID starting with 38e2f19a464a2f2f9d60ea157546f879ee6e670f10cadf94056461ca07855949 not found: ID does not exist" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.942875 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbb20043-1930-4c2d-b0d9-0cc7403cfc38-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.952960 4745 scope.go:117] "RemoveContainer" containerID="0ddc1787eb1e230a614cf715c9d31bc4b65fdee8e522ab0d1fb86425f72ad96d" Oct 10 13:43:12 crc kubenswrapper[4745]: I1010 13:43:12.992656 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cn7lb"] Oct 10 13:43:13 crc kubenswrapper[4745]: I1010 13:43:13.000321 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cn7lb"] Oct 10 13:43:14 crc kubenswrapper[4745]: I1010 13:43:14.034116 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-7cqks"] Oct 10 13:43:14 crc kubenswrapper[4745]: I1010 13:43:14.047235 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-7cqks"] Oct 10 13:43:14 crc kubenswrapper[4745]: I1010 13:43:14.760779 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c942887-aa59-417d-9ea5-e66cd98a1156" path="/var/lib/kubelet/pods/7c942887-aa59-417d-9ea5-e66cd98a1156/volumes" Oct 10 13:43:14 crc kubenswrapper[4745]: I1010 13:43:14.762362 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbb20043-1930-4c2d-b0d9-0cc7403cfc38" path="/var/lib/kubelet/pods/dbb20043-1930-4c2d-b0d9-0cc7403cfc38/volumes" Oct 10 13:43:16 crc kubenswrapper[4745]: I1010 13:43:16.025863 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-t5tfd"] Oct 10 13:43:16 crc kubenswrapper[4745]: I1010 13:43:16.033167 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-zw4m7"] Oct 10 13:43:16 crc kubenswrapper[4745]: I1010 13:43:16.041053 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-t5tfd"] Oct 10 13:43:16 crc kubenswrapper[4745]: I1010 13:43:16.048303 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-zw4m7"] Oct 10 13:43:16 crc kubenswrapper[4745]: I1010 13:43:16.759683 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="531008bf-6675-4b72-83a5-3aaabf54a133" path="/var/lib/kubelet/pods/531008bf-6675-4b72-83a5-3aaabf54a133/volumes" Oct 10 13:43:16 crc kubenswrapper[4745]: I1010 13:43:16.760412 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75ea46f9-2e99-4f8d-ac69-053ba3f2e71d" path="/var/lib/kubelet/pods/75ea46f9-2e99-4f8d-ac69-053ba3f2e71d/volumes" Oct 10 13:43:19 crc kubenswrapper[4745]: I1010 13:43:19.040038 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-nfdck"] Oct 10 13:43:19 crc kubenswrapper[4745]: I1010 13:43:19.046311 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-nfdck"] Oct 10 13:43:20 crc kubenswrapper[4745]: I1010 13:43:20.027969 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-bdv5b"] Oct 10 13:43:20 crc kubenswrapper[4745]: I1010 13:43:20.037517 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-bdv5b"] Oct 10 13:43:20 crc kubenswrapper[4745]: I1010 13:43:20.746051 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:43:20 crc kubenswrapper[4745]: E1010 13:43:20.746536 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:43:20 crc kubenswrapper[4745]: I1010 13:43:20.760092 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d439664-95e3-4c4d-b22b-a14242767a3a" path="/var/lib/kubelet/pods/2d439664-95e3-4c4d-b22b-a14242767a3a/volumes" Oct 10 13:43:20 crc kubenswrapper[4745]: I1010 13:43:20.762257 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dfdede3-b1b6-4630-afe1-0c2cdecf57f1" path="/var/lib/kubelet/pods/6dfdede3-b1b6-4630-afe1-0c2cdecf57f1/volumes" Oct 10 13:43:28 crc kubenswrapper[4745]: I1010 13:43:28.839380 4745 generic.go:334] "Generic (PLEG): container finished" podID="e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18" containerID="655adce735902766a08fa9aa5dab7a8bca060cd3f8c615446e0b86bb56e26e41" exitCode=0 Oct 10 13:43:28 crc kubenswrapper[4745]: I1010 13:43:28.839456 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" event={"ID":"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18","Type":"ContainerDied","Data":"655adce735902766a08fa9aa5dab7a8bca060cd3f8c615446e0b86bb56e26e41"} Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.312135 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.401033 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-inventory\") pod \"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18\" (UID: \"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18\") " Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.401480 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcpbc\" (UniqueName: \"kubernetes.io/projected/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-kube-api-access-qcpbc\") pod \"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18\" (UID: \"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18\") " Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.401519 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-ssh-key\") pod \"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18\" (UID: \"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18\") " Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.408749 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-kube-api-access-qcpbc" (OuterVolumeSpecName: "kube-api-access-qcpbc") pod "e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18" (UID: "e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18"). InnerVolumeSpecName "kube-api-access-qcpbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.435649 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18" (UID: "e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.436968 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-inventory" (OuterVolumeSpecName: "inventory") pod "e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18" (UID: "e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.504419 4745 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.504455 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcpbc\" (UniqueName: \"kubernetes.io/projected/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-kube-api-access-qcpbc\") on node \"crc\" DevicePath \"\"" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.504468 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.866820 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" event={"ID":"e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18","Type":"ContainerDied","Data":"61ac50aee2ce804a9036a05c70a44d26f5195f195197022378940008142e3578"} Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.866922 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61ac50aee2ce804a9036a05c70a44d26f5195f195197022378940008142e3578" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.866937 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.964725 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx"] Oct 10 13:43:30 crc kubenswrapper[4745]: E1010 13:43:30.965212 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.965237 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 10 13:43:30 crc kubenswrapper[4745]: E1010 13:43:30.965269 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbb20043-1930-4c2d-b0d9-0cc7403cfc38" containerName="extract-content" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.965279 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbb20043-1930-4c2d-b0d9-0cc7403cfc38" containerName="extract-content" Oct 10 13:43:30 crc kubenswrapper[4745]: E1010 13:43:30.965293 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbb20043-1930-4c2d-b0d9-0cc7403cfc38" containerName="registry-server" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.965301 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbb20043-1930-4c2d-b0d9-0cc7403cfc38" containerName="registry-server" Oct 10 13:43:30 crc kubenswrapper[4745]: E1010 13:43:30.965312 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbb20043-1930-4c2d-b0d9-0cc7403cfc38" containerName="extract-utilities" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.965320 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbb20043-1930-4c2d-b0d9-0cc7403cfc38" containerName="extract-utilities" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.965538 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.965563 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbb20043-1930-4c2d-b0d9-0cc7403cfc38" containerName="registry-server" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.966171 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.968178 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.968972 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.969343 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.971423 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:43:30 crc kubenswrapper[4745]: I1010 13:43:30.981136 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx"] Oct 10 13:43:31 crc kubenswrapper[4745]: I1010 13:43:31.032411 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-09e0-account-create-qsb99"] Oct 10 13:43:31 crc kubenswrapper[4745]: I1010 13:43:31.039335 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-09e0-account-create-qsb99"] Oct 10 13:43:31 crc kubenswrapper[4745]: I1010 13:43:31.116875 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/986809df-787f-46ec-b4ac-dc28eaf82cc7-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xprnx\" (UID: \"986809df-787f-46ec-b4ac-dc28eaf82cc7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" Oct 10 13:43:31 crc kubenswrapper[4745]: I1010 13:43:31.116932 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7z59\" (UniqueName: \"kubernetes.io/projected/986809df-787f-46ec-b4ac-dc28eaf82cc7-kube-api-access-j7z59\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xprnx\" (UID: \"986809df-787f-46ec-b4ac-dc28eaf82cc7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" Oct 10 13:43:31 crc kubenswrapper[4745]: I1010 13:43:31.117050 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/986809df-787f-46ec-b4ac-dc28eaf82cc7-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xprnx\" (UID: \"986809df-787f-46ec-b4ac-dc28eaf82cc7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" Oct 10 13:43:31 crc kubenswrapper[4745]: I1010 13:43:31.220046 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/986809df-787f-46ec-b4ac-dc28eaf82cc7-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xprnx\" (UID: \"986809df-787f-46ec-b4ac-dc28eaf82cc7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" Oct 10 13:43:31 crc kubenswrapper[4745]: I1010 13:43:31.220149 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7z59\" (UniqueName: \"kubernetes.io/projected/986809df-787f-46ec-b4ac-dc28eaf82cc7-kube-api-access-j7z59\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xprnx\" (UID: \"986809df-787f-46ec-b4ac-dc28eaf82cc7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" Oct 10 13:43:31 crc kubenswrapper[4745]: I1010 13:43:31.220430 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/986809df-787f-46ec-b4ac-dc28eaf82cc7-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xprnx\" (UID: \"986809df-787f-46ec-b4ac-dc28eaf82cc7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" Oct 10 13:43:31 crc kubenswrapper[4745]: I1010 13:43:31.224647 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/986809df-787f-46ec-b4ac-dc28eaf82cc7-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xprnx\" (UID: \"986809df-787f-46ec-b4ac-dc28eaf82cc7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" Oct 10 13:43:31 crc kubenswrapper[4745]: I1010 13:43:31.233492 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/986809df-787f-46ec-b4ac-dc28eaf82cc7-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xprnx\" (UID: \"986809df-787f-46ec-b4ac-dc28eaf82cc7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" Oct 10 13:43:31 crc kubenswrapper[4745]: I1010 13:43:31.244311 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7z59\" (UniqueName: \"kubernetes.io/projected/986809df-787f-46ec-b4ac-dc28eaf82cc7-kube-api-access-j7z59\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xprnx\" (UID: \"986809df-787f-46ec-b4ac-dc28eaf82cc7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" Oct 10 13:43:31 crc kubenswrapper[4745]: I1010 13:43:31.295055 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" Oct 10 13:43:31 crc kubenswrapper[4745]: I1010 13:43:31.838020 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx"] Oct 10 13:43:31 crc kubenswrapper[4745]: I1010 13:43:31.848403 4745 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 13:43:31 crc kubenswrapper[4745]: I1010 13:43:31.877409 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" event={"ID":"986809df-787f-46ec-b4ac-dc28eaf82cc7","Type":"ContainerStarted","Data":"c306d5fd2c910c6bbcb7a56be3472b82a68b6bc199fddfab0886b96003d01004"} Oct 10 13:43:32 crc kubenswrapper[4745]: I1010 13:43:32.747008 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:43:32 crc kubenswrapper[4745]: E1010 13:43:32.747647 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:43:32 crc kubenswrapper[4745]: I1010 13:43:32.762562 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="335e23ac-b3dc-4bc0-b7e1-505deb67af27" path="/var/lib/kubelet/pods/335e23ac-b3dc-4bc0-b7e1-505deb67af27/volumes" Oct 10 13:43:32 crc kubenswrapper[4745]: I1010 13:43:32.894321 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" event={"ID":"986809df-787f-46ec-b4ac-dc28eaf82cc7","Type":"ContainerStarted","Data":"431d1cfbe0dff4caff13100828e9ad2df781d6ed5104b7b289b570c848453441"} Oct 10 13:43:32 crc kubenswrapper[4745]: I1010 13:43:32.926556 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" podStartSLOduration=2.452967299 podStartE2EDuration="2.926533309s" podCreationTimestamp="2025-10-10 13:43:30 +0000 UTC" firstStartedPulling="2025-10-10 13:43:31.847590612 +0000 UTC m=+1525.745247415" lastFinishedPulling="2025-10-10 13:43:32.321156662 +0000 UTC m=+1526.218813425" observedRunningTime="2025-10-10 13:43:32.920126366 +0000 UTC m=+1526.817783169" watchObservedRunningTime="2025-10-10 13:43:32.926533309 +0000 UTC m=+1526.824190082" Oct 10 13:43:40 crc kubenswrapper[4745]: I1010 13:43:40.038538 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-98a4-account-create-r4qqn"] Oct 10 13:43:40 crc kubenswrapper[4745]: I1010 13:43:40.048366 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-c875-account-create-mrv46"] Oct 10 13:43:40 crc kubenswrapper[4745]: I1010 13:43:40.057703 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-98a4-account-create-r4qqn"] Oct 10 13:43:40 crc kubenswrapper[4745]: I1010 13:43:40.064159 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-c875-account-create-mrv46"] Oct 10 13:43:40 crc kubenswrapper[4745]: I1010 13:43:40.765393 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="235509a2-b516-432d-a681-9db08de1e7eb" path="/var/lib/kubelet/pods/235509a2-b516-432d-a681-9db08de1e7eb/volumes" Oct 10 13:43:40 crc kubenswrapper[4745]: I1010 13:43:40.766983 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a27645b6-74d3-4a15-ab07-c9767ef55413" path="/var/lib/kubelet/pods/a27645b6-74d3-4a15-ab07-c9767ef55413/volumes" Oct 10 13:43:47 crc kubenswrapper[4745]: I1010 13:43:47.745140 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:43:47 crc kubenswrapper[4745]: E1010 13:43:47.746004 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:43:51 crc kubenswrapper[4745]: I1010 13:43:51.044601 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-p9wdk"] Oct 10 13:43:51 crc kubenswrapper[4745]: I1010 13:43:51.067322 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-p9wdk"] Oct 10 13:43:52 crc kubenswrapper[4745]: I1010 13:43:52.760026 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28786a80-f298-4618-8c54-5f943d96eff9" path="/var/lib/kubelet/pods/28786a80-f298-4618-8c54-5f943d96eff9/volumes" Oct 10 13:43:59 crc kubenswrapper[4745]: I1010 13:43:59.745689 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:43:59 crc kubenswrapper[4745]: E1010 13:43:59.746718 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:44:08 crc kubenswrapper[4745]: I1010 13:44:08.055303 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-krvls"] Oct 10 13:44:08 crc kubenswrapper[4745]: I1010 13:44:08.063816 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-krvls"] Oct 10 13:44:08 crc kubenswrapper[4745]: I1010 13:44:08.760111 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45009f2e-d0d5-4a20-b62c-d5d0534e1551" path="/var/lib/kubelet/pods/45009f2e-d0d5-4a20-b62c-d5d0534e1551/volumes" Oct 10 13:44:12 crc kubenswrapper[4745]: I1010 13:44:12.745071 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:44:12 crc kubenswrapper[4745]: E1010 13:44:12.745747 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:44:13 crc kubenswrapper[4745]: I1010 13:44:13.136190 4745 scope.go:117] "RemoveContainer" containerID="9497ef0ece50c843aef82bd37fb30c7a6e552c24bf36a6a35253efce4286198e" Oct 10 13:44:13 crc kubenswrapper[4745]: I1010 13:44:13.164437 4745 scope.go:117] "RemoveContainer" containerID="45e243e61dc30b54e185fd498af873db27c34aba4f03cd1ae9945b80fe22815a" Oct 10 13:44:13 crc kubenswrapper[4745]: I1010 13:44:13.240815 4745 scope.go:117] "RemoveContainer" containerID="d8e4e4690c3e55f3e93c6501f969fa2a4ea9030d0515f2fb876ab1921d0fb1e4" Oct 10 13:44:13 crc kubenswrapper[4745]: I1010 13:44:13.274149 4745 scope.go:117] "RemoveContainer" containerID="29be1db3925cfbfc891675f55b1fcbd20948a76fa309b906903dbecfe3f9ba1a" Oct 10 13:44:13 crc kubenswrapper[4745]: I1010 13:44:13.318602 4745 scope.go:117] "RemoveContainer" containerID="b76bccde80d9fa6432d17a2c1afe41626d6403bf036a993e9072ce4023acd22b" Oct 10 13:44:13 crc kubenswrapper[4745]: I1010 13:44:13.368336 4745 scope.go:117] "RemoveContainer" containerID="2feaf091557339dc922e1a4b1bd65e94052297f597c62ab49422faa9301c2315" Oct 10 13:44:13 crc kubenswrapper[4745]: I1010 13:44:13.445971 4745 scope.go:117] "RemoveContainer" containerID="00bb0bf4c37ec198785551594973932ef355f5ed17da60e5fae2c842a5f44001" Oct 10 13:44:13 crc kubenswrapper[4745]: I1010 13:44:13.477857 4745 scope.go:117] "RemoveContainer" containerID="45effcc901700734d808a25087a434d0114d0bb624d3707963710d7581f4b9f9" Oct 10 13:44:13 crc kubenswrapper[4745]: I1010 13:44:13.499944 4745 scope.go:117] "RemoveContainer" containerID="8f2c29aec06f8555ebac52308b3893ce31a90233c10d431b1255d3755d2ea170" Oct 10 13:44:13 crc kubenswrapper[4745]: I1010 13:44:13.523659 4745 scope.go:117] "RemoveContainer" containerID="50b424a7887fcd3389e8d086c4b8417714a799fcf3d1110d2ab5273a60a5ec18" Oct 10 13:44:19 crc kubenswrapper[4745]: I1010 13:44:19.036154 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-c2vrp"] Oct 10 13:44:19 crc kubenswrapper[4745]: I1010 13:44:19.043979 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-dd6lk"] Oct 10 13:44:19 crc kubenswrapper[4745]: I1010 13:44:19.052837 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-dd6lk"] Oct 10 13:44:19 crc kubenswrapper[4745]: I1010 13:44:19.061508 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-c2vrp"] Oct 10 13:44:20 crc kubenswrapper[4745]: I1010 13:44:20.761767 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60ccdbf7-fc2d-4292-8ff6-d388ccda2b58" path="/var/lib/kubelet/pods/60ccdbf7-fc2d-4292-8ff6-d388ccda2b58/volumes" Oct 10 13:44:20 crc kubenswrapper[4745]: I1010 13:44:20.763089 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cbce4a4-b5ac-481c-87a7-47bc07e3361a" path="/var/lib/kubelet/pods/7cbce4a4-b5ac-481c-87a7-47bc07e3361a/volumes" Oct 10 13:44:24 crc kubenswrapper[4745]: I1010 13:44:24.745235 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:44:24 crc kubenswrapper[4745]: E1010 13:44:24.745970 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:44:30 crc kubenswrapper[4745]: I1010 13:44:30.058823 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-vskzt"] Oct 10 13:44:30 crc kubenswrapper[4745]: I1010 13:44:30.072835 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-vskzt"] Oct 10 13:44:30 crc kubenswrapper[4745]: I1010 13:44:30.758749 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e86aa13-6741-4a98-88f7-1fec0870468f" path="/var/lib/kubelet/pods/7e86aa13-6741-4a98-88f7-1fec0870468f/volumes" Oct 10 13:44:38 crc kubenswrapper[4745]: I1010 13:44:38.745135 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:44:38 crc kubenswrapper[4745]: E1010 13:44:38.745912 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:44:40 crc kubenswrapper[4745]: I1010 13:44:40.041196 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-z2v8n"] Oct 10 13:44:40 crc kubenswrapper[4745]: I1010 13:44:40.055968 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-z2v8n"] Oct 10 13:44:40 crc kubenswrapper[4745]: I1010 13:44:40.764113 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc3cdd18-7362-4610-8f9b-7444bd45b09c" path="/var/lib/kubelet/pods/cc3cdd18-7362-4610-8f9b-7444bd45b09c/volumes" Oct 10 13:44:43 crc kubenswrapper[4745]: I1010 13:44:43.025790 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-h8ngs"] Oct 10 13:44:43 crc kubenswrapper[4745]: I1010 13:44:43.032935 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-xc9ls"] Oct 10 13:44:43 crc kubenswrapper[4745]: I1010 13:44:43.039563 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-h8ngs"] Oct 10 13:44:43 crc kubenswrapper[4745]: I1010 13:44:43.051133 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-xc9ls"] Oct 10 13:44:44 crc kubenswrapper[4745]: I1010 13:44:44.630442 4745 generic.go:334] "Generic (PLEG): container finished" podID="986809df-787f-46ec-b4ac-dc28eaf82cc7" containerID="431d1cfbe0dff4caff13100828e9ad2df781d6ed5104b7b289b570c848453441" exitCode=0 Oct 10 13:44:44 crc kubenswrapper[4745]: I1010 13:44:44.630480 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" event={"ID":"986809df-787f-46ec-b4ac-dc28eaf82cc7","Type":"ContainerDied","Data":"431d1cfbe0dff4caff13100828e9ad2df781d6ed5104b7b289b570c848453441"} Oct 10 13:44:44 crc kubenswrapper[4745]: I1010 13:44:44.755893 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28904cbe-09fb-499a-8d88-23ac50034c90" path="/var/lib/kubelet/pods/28904cbe-09fb-499a-8d88-23ac50034c90/volumes" Oct 10 13:44:44 crc kubenswrapper[4745]: I1010 13:44:44.757221 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3771609-ba86-4413-8cbd-3d266cd0380d" path="/var/lib/kubelet/pods/a3771609-ba86-4413-8cbd-3d266cd0380d/volumes" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.077194 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.196188 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/986809df-787f-46ec-b4ac-dc28eaf82cc7-inventory\") pod \"986809df-787f-46ec-b4ac-dc28eaf82cc7\" (UID: \"986809df-787f-46ec-b4ac-dc28eaf82cc7\") " Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.196299 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/986809df-787f-46ec-b4ac-dc28eaf82cc7-ssh-key\") pod \"986809df-787f-46ec-b4ac-dc28eaf82cc7\" (UID: \"986809df-787f-46ec-b4ac-dc28eaf82cc7\") " Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.196382 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7z59\" (UniqueName: \"kubernetes.io/projected/986809df-787f-46ec-b4ac-dc28eaf82cc7-kube-api-access-j7z59\") pod \"986809df-787f-46ec-b4ac-dc28eaf82cc7\" (UID: \"986809df-787f-46ec-b4ac-dc28eaf82cc7\") " Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.203014 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/986809df-787f-46ec-b4ac-dc28eaf82cc7-kube-api-access-j7z59" (OuterVolumeSpecName: "kube-api-access-j7z59") pod "986809df-787f-46ec-b4ac-dc28eaf82cc7" (UID: "986809df-787f-46ec-b4ac-dc28eaf82cc7"). InnerVolumeSpecName "kube-api-access-j7z59". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.231239 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/986809df-787f-46ec-b4ac-dc28eaf82cc7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "986809df-787f-46ec-b4ac-dc28eaf82cc7" (UID: "986809df-787f-46ec-b4ac-dc28eaf82cc7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.240528 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/986809df-787f-46ec-b4ac-dc28eaf82cc7-inventory" (OuterVolumeSpecName: "inventory") pod "986809df-787f-46ec-b4ac-dc28eaf82cc7" (UID: "986809df-787f-46ec-b4ac-dc28eaf82cc7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.298370 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7z59\" (UniqueName: \"kubernetes.io/projected/986809df-787f-46ec-b4ac-dc28eaf82cc7-kube-api-access-j7z59\") on node \"crc\" DevicePath \"\"" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.298406 4745 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/986809df-787f-46ec-b4ac-dc28eaf82cc7-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.298417 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/986809df-787f-46ec-b4ac-dc28eaf82cc7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.651721 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" event={"ID":"986809df-787f-46ec-b4ac-dc28eaf82cc7","Type":"ContainerDied","Data":"c306d5fd2c910c6bbcb7a56be3472b82a68b6bc199fddfab0886b96003d01004"} Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.651818 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c306d5fd2c910c6bbcb7a56be3472b82a68b6bc199fddfab0886b96003d01004" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.651825 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xprnx" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.794773 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582"] Oct 10 13:44:46 crc kubenswrapper[4745]: E1010 13:44:46.795174 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="986809df-787f-46ec-b4ac-dc28eaf82cc7" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.795196 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="986809df-787f-46ec-b4ac-dc28eaf82cc7" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.795410 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="986809df-787f-46ec-b4ac-dc28eaf82cc7" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.796088 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.803556 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.803846 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.803851 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.804099 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.813394 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582"] Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.912568 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b9582\" (UID: \"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.912618 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn9s4\" (UniqueName: \"kubernetes.io/projected/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-kube-api-access-zn9s4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b9582\" (UID: \"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" Oct 10 13:44:46 crc kubenswrapper[4745]: I1010 13:44:46.912799 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b9582\" (UID: \"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" Oct 10 13:44:47 crc kubenswrapper[4745]: I1010 13:44:47.014721 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b9582\" (UID: \"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" Oct 10 13:44:47 crc kubenswrapper[4745]: I1010 13:44:47.014990 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b9582\" (UID: \"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" Oct 10 13:44:47 crc kubenswrapper[4745]: I1010 13:44:47.015032 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn9s4\" (UniqueName: \"kubernetes.io/projected/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-kube-api-access-zn9s4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b9582\" (UID: \"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" Oct 10 13:44:47 crc kubenswrapper[4745]: I1010 13:44:47.022415 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b9582\" (UID: \"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" Oct 10 13:44:47 crc kubenswrapper[4745]: I1010 13:44:47.026307 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b9582\" (UID: \"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" Oct 10 13:44:47 crc kubenswrapper[4745]: I1010 13:44:47.042850 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn9s4\" (UniqueName: \"kubernetes.io/projected/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-kube-api-access-zn9s4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-b9582\" (UID: \"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" Oct 10 13:44:47 crc kubenswrapper[4745]: I1010 13:44:47.129965 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" Oct 10 13:44:47 crc kubenswrapper[4745]: I1010 13:44:47.704829 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582"] Oct 10 13:44:48 crc kubenswrapper[4745]: I1010 13:44:48.678278 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" event={"ID":"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2","Type":"ContainerStarted","Data":"ce50efc588913e178f4ba6c602573cc25f62ebfc8c4bebfa42ee1716284729f9"} Oct 10 13:44:48 crc kubenswrapper[4745]: I1010 13:44:48.678369 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" event={"ID":"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2","Type":"ContainerStarted","Data":"03113fcaee789377115ceb518855b03a68084ac535748b1c4c62e82823e20222"} Oct 10 13:44:48 crc kubenswrapper[4745]: I1010 13:44:48.708254 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" podStartSLOduration=2.043334608 podStartE2EDuration="2.708231397s" podCreationTimestamp="2025-10-10 13:44:46 +0000 UTC" firstStartedPulling="2025-10-10 13:44:47.715125371 +0000 UTC m=+1601.612782134" lastFinishedPulling="2025-10-10 13:44:48.38002213 +0000 UTC m=+1602.277678923" observedRunningTime="2025-10-10 13:44:48.697873964 +0000 UTC m=+1602.595530767" watchObservedRunningTime="2025-10-10 13:44:48.708231397 +0000 UTC m=+1602.605888180" Oct 10 13:44:53 crc kubenswrapper[4745]: I1010 13:44:53.725513 4745 generic.go:334] "Generic (PLEG): container finished" podID="e4f15e34-6886-448d-9ed1-7a5cbaf0fda2" containerID="ce50efc588913e178f4ba6c602573cc25f62ebfc8c4bebfa42ee1716284729f9" exitCode=0 Oct 10 13:44:53 crc kubenswrapper[4745]: I1010 13:44:53.725617 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" event={"ID":"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2","Type":"ContainerDied","Data":"ce50efc588913e178f4ba6c602573cc25f62ebfc8c4bebfa42ee1716284729f9"} Oct 10 13:44:53 crc kubenswrapper[4745]: I1010 13:44:53.745378 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:44:53 crc kubenswrapper[4745]: E1010 13:44:53.745716 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:44:54 crc kubenswrapper[4745]: I1010 13:44:54.044911 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-618f-account-create-vhdv7"] Oct 10 13:44:54 crc kubenswrapper[4745]: I1010 13:44:54.061241 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-3a63-account-create-zprfs"] Oct 10 13:44:54 crc kubenswrapper[4745]: I1010 13:44:54.073093 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-3756-account-create-4dd9c"] Oct 10 13:44:54 crc kubenswrapper[4745]: I1010 13:44:54.081300 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-618f-account-create-vhdv7"] Oct 10 13:44:54 crc kubenswrapper[4745]: I1010 13:44:54.092971 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-3756-account-create-4dd9c"] Oct 10 13:44:54 crc kubenswrapper[4745]: I1010 13:44:54.103969 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-3a63-account-create-zprfs"] Oct 10 13:44:54 crc kubenswrapper[4745]: I1010 13:44:54.769621 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f7a4122-1442-4714-8556-da06a209febe" path="/var/lib/kubelet/pods/9f7a4122-1442-4714-8556-da06a209febe/volumes" Oct 10 13:44:54 crc kubenswrapper[4745]: I1010 13:44:54.771153 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bff51ec2-99e6-447c-9a70-47da73fc9b5d" path="/var/lib/kubelet/pods/bff51ec2-99e6-447c-9a70-47da73fc9b5d/volumes" Oct 10 13:44:54 crc kubenswrapper[4745]: I1010 13:44:54.772532 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d41388d5-c639-462f-a5ea-08a27f382511" path="/var/lib/kubelet/pods/d41388d5-c639-462f-a5ea-08a27f382511/volumes" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.161539 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.282223 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn9s4\" (UniqueName: \"kubernetes.io/projected/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-kube-api-access-zn9s4\") pod \"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2\" (UID: \"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2\") " Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.282605 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-ssh-key\") pod \"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2\" (UID: \"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2\") " Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.282861 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-inventory\") pod \"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2\" (UID: \"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2\") " Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.289668 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-kube-api-access-zn9s4" (OuterVolumeSpecName: "kube-api-access-zn9s4") pod "e4f15e34-6886-448d-9ed1-7a5cbaf0fda2" (UID: "e4f15e34-6886-448d-9ed1-7a5cbaf0fda2"). InnerVolumeSpecName "kube-api-access-zn9s4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.311083 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e4f15e34-6886-448d-9ed1-7a5cbaf0fda2" (UID: "e4f15e34-6886-448d-9ed1-7a5cbaf0fda2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.314935 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-inventory" (OuterVolumeSpecName: "inventory") pod "e4f15e34-6886-448d-9ed1-7a5cbaf0fda2" (UID: "e4f15e34-6886-448d-9ed1-7a5cbaf0fda2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.385335 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn9s4\" (UniqueName: \"kubernetes.io/projected/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-kube-api-access-zn9s4\") on node \"crc\" DevicePath \"\"" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.385375 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.385387 4745 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4f15e34-6886-448d-9ed1-7a5cbaf0fda2-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.752872 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" event={"ID":"e4f15e34-6886-448d-9ed1-7a5cbaf0fda2","Type":"ContainerDied","Data":"03113fcaee789377115ceb518855b03a68084ac535748b1c4c62e82823e20222"} Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.753426 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03113fcaee789377115ceb518855b03a68084ac535748b1c4c62e82823e20222" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.753532 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-b9582" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.865551 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg"] Oct 10 13:44:55 crc kubenswrapper[4745]: E1010 13:44:55.866208 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f15e34-6886-448d-9ed1-7a5cbaf0fda2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.866233 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f15e34-6886-448d-9ed1-7a5cbaf0fda2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.866611 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4f15e34-6886-448d-9ed1-7a5cbaf0fda2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.867689 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.869557 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.870615 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.871786 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.871799 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:44:55 crc kubenswrapper[4745]: I1010 13:44:55.887594 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg"] Oct 10 13:44:56 crc kubenswrapper[4745]: I1010 13:44:56.001614 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/292290b6-c1e2-4b2b-955b-699aa826627d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hflsg\" (UID: \"292290b6-c1e2-4b2b-955b-699aa826627d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" Oct 10 13:44:56 crc kubenswrapper[4745]: I1010 13:44:56.001924 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/292290b6-c1e2-4b2b-955b-699aa826627d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hflsg\" (UID: \"292290b6-c1e2-4b2b-955b-699aa826627d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" Oct 10 13:44:56 crc kubenswrapper[4745]: I1010 13:44:56.002128 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8vkq\" (UniqueName: \"kubernetes.io/projected/292290b6-c1e2-4b2b-955b-699aa826627d-kube-api-access-x8vkq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hflsg\" (UID: \"292290b6-c1e2-4b2b-955b-699aa826627d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" Oct 10 13:44:56 crc kubenswrapper[4745]: I1010 13:44:56.104593 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/292290b6-c1e2-4b2b-955b-699aa826627d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hflsg\" (UID: \"292290b6-c1e2-4b2b-955b-699aa826627d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" Oct 10 13:44:56 crc kubenswrapper[4745]: I1010 13:44:56.105099 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/292290b6-c1e2-4b2b-955b-699aa826627d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hflsg\" (UID: \"292290b6-c1e2-4b2b-955b-699aa826627d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" Oct 10 13:44:56 crc kubenswrapper[4745]: I1010 13:44:56.105487 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8vkq\" (UniqueName: \"kubernetes.io/projected/292290b6-c1e2-4b2b-955b-699aa826627d-kube-api-access-x8vkq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hflsg\" (UID: \"292290b6-c1e2-4b2b-955b-699aa826627d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" Oct 10 13:44:56 crc kubenswrapper[4745]: I1010 13:44:56.112992 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/292290b6-c1e2-4b2b-955b-699aa826627d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hflsg\" (UID: \"292290b6-c1e2-4b2b-955b-699aa826627d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" Oct 10 13:44:56 crc kubenswrapper[4745]: I1010 13:44:56.113359 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/292290b6-c1e2-4b2b-955b-699aa826627d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hflsg\" (UID: \"292290b6-c1e2-4b2b-955b-699aa826627d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" Oct 10 13:44:56 crc kubenswrapper[4745]: I1010 13:44:56.125838 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8vkq\" (UniqueName: \"kubernetes.io/projected/292290b6-c1e2-4b2b-955b-699aa826627d-kube-api-access-x8vkq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hflsg\" (UID: \"292290b6-c1e2-4b2b-955b-699aa826627d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" Oct 10 13:44:56 crc kubenswrapper[4745]: I1010 13:44:56.205172 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" Oct 10 13:44:56 crc kubenswrapper[4745]: W1010 13:44:56.760760 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod292290b6_c1e2_4b2b_955b_699aa826627d.slice/crio-a540d416ca52b002129f20694a9681407755be329615d8eca569f02ec865e979 WatchSource:0}: Error finding container a540d416ca52b002129f20694a9681407755be329615d8eca569f02ec865e979: Status 404 returned error can't find the container with id a540d416ca52b002129f20694a9681407755be329615d8eca569f02ec865e979 Oct 10 13:44:56 crc kubenswrapper[4745]: I1010 13:44:56.780531 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg"] Oct 10 13:44:57 crc kubenswrapper[4745]: I1010 13:44:57.781881 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" event={"ID":"292290b6-c1e2-4b2b-955b-699aa826627d","Type":"ContainerStarted","Data":"bfe9ded5fe41ed255774418ffa2edb1f635d903a7c9a208532e8e401ef1b2ba9"} Oct 10 13:44:57 crc kubenswrapper[4745]: I1010 13:44:57.782239 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" event={"ID":"292290b6-c1e2-4b2b-955b-699aa826627d","Type":"ContainerStarted","Data":"a540d416ca52b002129f20694a9681407755be329615d8eca569f02ec865e979"} Oct 10 13:44:57 crc kubenswrapper[4745]: I1010 13:44:57.802201 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" podStartSLOduration=2.253314986 podStartE2EDuration="2.802174138s" podCreationTimestamp="2025-10-10 13:44:55 +0000 UTC" firstStartedPulling="2025-10-10 13:44:56.77160136 +0000 UTC m=+1610.669258133" lastFinishedPulling="2025-10-10 13:44:57.320460512 +0000 UTC m=+1611.218117285" observedRunningTime="2025-10-10 13:44:57.800886566 +0000 UTC m=+1611.698543329" watchObservedRunningTime="2025-10-10 13:44:57.802174138 +0000 UTC m=+1611.699830911" Oct 10 13:45:00 crc kubenswrapper[4745]: I1010 13:45:00.144305 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb"] Oct 10 13:45:00 crc kubenswrapper[4745]: I1010 13:45:00.147280 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb" Oct 10 13:45:00 crc kubenswrapper[4745]: I1010 13:45:00.150704 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 13:45:00 crc kubenswrapper[4745]: I1010 13:45:00.151462 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 13:45:00 crc kubenswrapper[4745]: I1010 13:45:00.164571 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb"] Oct 10 13:45:00 crc kubenswrapper[4745]: I1010 13:45:00.193295 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pksk\" (UniqueName: \"kubernetes.io/projected/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-kube-api-access-4pksk\") pod \"collect-profiles-29335065-mbbrb\" (UID: \"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb" Oct 10 13:45:00 crc kubenswrapper[4745]: I1010 13:45:00.193398 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-secret-volume\") pod \"collect-profiles-29335065-mbbrb\" (UID: \"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb" Oct 10 13:45:00 crc kubenswrapper[4745]: I1010 13:45:00.193604 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-config-volume\") pod \"collect-profiles-29335065-mbbrb\" (UID: \"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb" Oct 10 13:45:00 crc kubenswrapper[4745]: I1010 13:45:00.296182 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pksk\" (UniqueName: \"kubernetes.io/projected/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-kube-api-access-4pksk\") pod \"collect-profiles-29335065-mbbrb\" (UID: \"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb" Oct 10 13:45:00 crc kubenswrapper[4745]: I1010 13:45:00.296503 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-secret-volume\") pod \"collect-profiles-29335065-mbbrb\" (UID: \"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb" Oct 10 13:45:00 crc kubenswrapper[4745]: I1010 13:45:00.296695 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-config-volume\") pod \"collect-profiles-29335065-mbbrb\" (UID: \"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb" Oct 10 13:45:00 crc kubenswrapper[4745]: I1010 13:45:00.298069 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-config-volume\") pod \"collect-profiles-29335065-mbbrb\" (UID: \"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb" Oct 10 13:45:00 crc kubenswrapper[4745]: I1010 13:45:00.302169 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-secret-volume\") pod \"collect-profiles-29335065-mbbrb\" (UID: \"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb" Oct 10 13:45:00 crc kubenswrapper[4745]: I1010 13:45:00.316997 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pksk\" (UniqueName: \"kubernetes.io/projected/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-kube-api-access-4pksk\") pod \"collect-profiles-29335065-mbbrb\" (UID: \"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb" Oct 10 13:45:00 crc kubenswrapper[4745]: I1010 13:45:00.488675 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb" Oct 10 13:45:00 crc kubenswrapper[4745]: I1010 13:45:00.970582 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb"] Oct 10 13:45:01 crc kubenswrapper[4745]: I1010 13:45:01.819142 4745 generic.go:334] "Generic (PLEG): container finished" podID="ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e" containerID="ef0c07e993fa175f34a7b85bacbfc1f356e0e3cb90a006c218f254bed23f91f7" exitCode=0 Oct 10 13:45:01 crc kubenswrapper[4745]: I1010 13:45:01.819186 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb" event={"ID":"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e","Type":"ContainerDied","Data":"ef0c07e993fa175f34a7b85bacbfc1f356e0e3cb90a006c218f254bed23f91f7"} Oct 10 13:45:01 crc kubenswrapper[4745]: I1010 13:45:01.819210 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb" event={"ID":"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e","Type":"ContainerStarted","Data":"5d5f07eef4fe3360c4d706e65158afdd6163c2a524a72f290e0dcb1c09239251"} Oct 10 13:45:03 crc kubenswrapper[4745]: I1010 13:45:03.124233 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb" Oct 10 13:45:03 crc kubenswrapper[4745]: I1010 13:45:03.259606 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pksk\" (UniqueName: \"kubernetes.io/projected/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-kube-api-access-4pksk\") pod \"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e\" (UID: \"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e\") " Oct 10 13:45:03 crc kubenswrapper[4745]: I1010 13:45:03.259751 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-config-volume\") pod \"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e\" (UID: \"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e\") " Oct 10 13:45:03 crc kubenswrapper[4745]: I1010 13:45:03.259839 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-secret-volume\") pod \"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e\" (UID: \"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e\") " Oct 10 13:45:03 crc kubenswrapper[4745]: I1010 13:45:03.260465 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-config-volume" (OuterVolumeSpecName: "config-volume") pod "ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e" (UID: "ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:45:03 crc kubenswrapper[4745]: I1010 13:45:03.265194 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e" (UID: "ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:45:03 crc kubenswrapper[4745]: I1010 13:45:03.265620 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-kube-api-access-4pksk" (OuterVolumeSpecName: "kube-api-access-4pksk") pod "ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e" (UID: "ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e"). InnerVolumeSpecName "kube-api-access-4pksk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:45:03 crc kubenswrapper[4745]: I1010 13:45:03.361971 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pksk\" (UniqueName: \"kubernetes.io/projected/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-kube-api-access-4pksk\") on node \"crc\" DevicePath \"\"" Oct 10 13:45:03 crc kubenswrapper[4745]: I1010 13:45:03.362004 4745 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 13:45:03 crc kubenswrapper[4745]: I1010 13:45:03.362014 4745 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 13:45:03 crc kubenswrapper[4745]: I1010 13:45:03.839445 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb" event={"ID":"ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e","Type":"ContainerDied","Data":"5d5f07eef4fe3360c4d706e65158afdd6163c2a524a72f290e0dcb1c09239251"} Oct 10 13:45:03 crc kubenswrapper[4745]: I1010 13:45:03.839489 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d5f07eef4fe3360c4d706e65158afdd6163c2a524a72f290e0dcb1c09239251" Oct 10 13:45:03 crc kubenswrapper[4745]: I1010 13:45:03.839495 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335065-mbbrb" Oct 10 13:45:04 crc kubenswrapper[4745]: I1010 13:45:04.744459 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:45:04 crc kubenswrapper[4745]: E1010 13:45:04.745100 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:45:13 crc kubenswrapper[4745]: I1010 13:45:13.771439 4745 scope.go:117] "RemoveContainer" containerID="19b3dcb6e0090f97c5a0c0bbf520c77933b0c453d11c50ece85fea9b057f4eb2" Oct 10 13:45:13 crc kubenswrapper[4745]: I1010 13:45:13.812429 4745 scope.go:117] "RemoveContainer" containerID="e674aae9e59433329468c0bf150bafab118433d833fc27a5b1f5fb24a51598a1" Oct 10 13:45:13 crc kubenswrapper[4745]: I1010 13:45:13.897555 4745 scope.go:117] "RemoveContainer" containerID="262bb16c312dd05eb8a87ac07d05f484e890cc5a2a2441b648b10260e103fa6e" Oct 10 13:45:13 crc kubenswrapper[4745]: I1010 13:45:13.947110 4745 scope.go:117] "RemoveContainer" containerID="61accf956903ff3eed21347a235a981ff594f9680eb014b7f8c21fa2a0e96599" Oct 10 13:45:13 crc kubenswrapper[4745]: I1010 13:45:13.983504 4745 scope.go:117] "RemoveContainer" containerID="5cee0587bc53b5ae5ba9b9d87902512f80337b576beb7c07a230a232d8f6d114" Oct 10 13:45:14 crc kubenswrapper[4745]: I1010 13:45:14.025296 4745 scope.go:117] "RemoveContainer" containerID="c6590cccecf9b423f770e4917fe5c07b52770199f5d0b44840bfe66534fb9e90" Oct 10 13:45:14 crc kubenswrapper[4745]: I1010 13:45:14.072654 4745 scope.go:117] "RemoveContainer" containerID="63a78cd681c0ab49c65e8b0702691600a179cf2b4f4e6340c54a309ca98cddd9" Oct 10 13:45:14 crc kubenswrapper[4745]: I1010 13:45:14.102966 4745 scope.go:117] "RemoveContainer" containerID="f1531086e8b678df707da20bde70fd192a60eb1893dd4bc552ae427f3a109748" Oct 10 13:45:14 crc kubenswrapper[4745]: I1010 13:45:14.135635 4745 scope.go:117] "RemoveContainer" containerID="94cec6aba343eb4bb33675324772ba9f235d6b31718fcf8704c4fd2f5e178063" Oct 10 13:45:18 crc kubenswrapper[4745]: I1010 13:45:18.745271 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:45:18 crc kubenswrapper[4745]: E1010 13:45:18.746692 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:45:20 crc kubenswrapper[4745]: I1010 13:45:20.058359 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wp869"] Oct 10 13:45:20 crc kubenswrapper[4745]: I1010 13:45:20.065387 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wp869"] Oct 10 13:45:20 crc kubenswrapper[4745]: I1010 13:45:20.769530 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="907a33f4-60a8-44b7-8b1d-5fb979a84e62" path="/var/lib/kubelet/pods/907a33f4-60a8-44b7-8b1d-5fb979a84e62/volumes" Oct 10 13:45:33 crc kubenswrapper[4745]: I1010 13:45:33.745924 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:45:33 crc kubenswrapper[4745]: E1010 13:45:33.747027 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:45:38 crc kubenswrapper[4745]: I1010 13:45:38.038097 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-8t6r5"] Oct 10 13:45:38 crc kubenswrapper[4745]: I1010 13:45:38.045685 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-8t6r5"] Oct 10 13:45:38 crc kubenswrapper[4745]: I1010 13:45:38.755127 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50933493-e7d7-4fc2-93bc-69173739fe58" path="/var/lib/kubelet/pods/50933493-e7d7-4fc2-93bc-69173739fe58/volumes" Oct 10 13:45:39 crc kubenswrapper[4745]: I1010 13:45:39.053096 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6xlzw"] Oct 10 13:45:39 crc kubenswrapper[4745]: I1010 13:45:39.066216 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6xlzw"] Oct 10 13:45:39 crc kubenswrapper[4745]: I1010 13:45:39.219295 4745 generic.go:334] "Generic (PLEG): container finished" podID="292290b6-c1e2-4b2b-955b-699aa826627d" containerID="bfe9ded5fe41ed255774418ffa2edb1f635d903a7c9a208532e8e401ef1b2ba9" exitCode=0 Oct 10 13:45:39 crc kubenswrapper[4745]: I1010 13:45:39.219346 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" event={"ID":"292290b6-c1e2-4b2b-955b-699aa826627d","Type":"ContainerDied","Data":"bfe9ded5fe41ed255774418ffa2edb1f635d903a7c9a208532e8e401ef1b2ba9"} Oct 10 13:45:40 crc kubenswrapper[4745]: I1010 13:45:40.680449 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" Oct 10 13:45:40 crc kubenswrapper[4745]: I1010 13:45:40.763031 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="642aa23f-361f-4bc5-b463-15a104a8d7d7" path="/var/lib/kubelet/pods/642aa23f-361f-4bc5-b463-15a104a8d7d7/volumes" Oct 10 13:45:40 crc kubenswrapper[4745]: I1010 13:45:40.771696 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8vkq\" (UniqueName: \"kubernetes.io/projected/292290b6-c1e2-4b2b-955b-699aa826627d-kube-api-access-x8vkq\") pod \"292290b6-c1e2-4b2b-955b-699aa826627d\" (UID: \"292290b6-c1e2-4b2b-955b-699aa826627d\") " Oct 10 13:45:40 crc kubenswrapper[4745]: I1010 13:45:40.771891 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/292290b6-c1e2-4b2b-955b-699aa826627d-ssh-key\") pod \"292290b6-c1e2-4b2b-955b-699aa826627d\" (UID: \"292290b6-c1e2-4b2b-955b-699aa826627d\") " Oct 10 13:45:40 crc kubenswrapper[4745]: I1010 13:45:40.772026 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/292290b6-c1e2-4b2b-955b-699aa826627d-inventory\") pod \"292290b6-c1e2-4b2b-955b-699aa826627d\" (UID: \"292290b6-c1e2-4b2b-955b-699aa826627d\") " Oct 10 13:45:40 crc kubenswrapper[4745]: I1010 13:45:40.781128 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/292290b6-c1e2-4b2b-955b-699aa826627d-kube-api-access-x8vkq" (OuterVolumeSpecName: "kube-api-access-x8vkq") pod "292290b6-c1e2-4b2b-955b-699aa826627d" (UID: "292290b6-c1e2-4b2b-955b-699aa826627d"). InnerVolumeSpecName "kube-api-access-x8vkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:45:40 crc kubenswrapper[4745]: I1010 13:45:40.808686 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/292290b6-c1e2-4b2b-955b-699aa826627d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "292290b6-c1e2-4b2b-955b-699aa826627d" (UID: "292290b6-c1e2-4b2b-955b-699aa826627d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:45:40 crc kubenswrapper[4745]: I1010 13:45:40.817817 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/292290b6-c1e2-4b2b-955b-699aa826627d-inventory" (OuterVolumeSpecName: "inventory") pod "292290b6-c1e2-4b2b-955b-699aa826627d" (UID: "292290b6-c1e2-4b2b-955b-699aa826627d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:45:40 crc kubenswrapper[4745]: I1010 13:45:40.875443 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8vkq\" (UniqueName: \"kubernetes.io/projected/292290b6-c1e2-4b2b-955b-699aa826627d-kube-api-access-x8vkq\") on node \"crc\" DevicePath \"\"" Oct 10 13:45:40 crc kubenswrapper[4745]: I1010 13:45:40.875701 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/292290b6-c1e2-4b2b-955b-699aa826627d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:45:40 crc kubenswrapper[4745]: I1010 13:45:40.875853 4745 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/292290b6-c1e2-4b2b-955b-699aa826627d-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.241725 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" event={"ID":"292290b6-c1e2-4b2b-955b-699aa826627d","Type":"ContainerDied","Data":"a540d416ca52b002129f20694a9681407755be329615d8eca569f02ec865e979"} Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.241995 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a540d416ca52b002129f20694a9681407755be329615d8eca569f02ec865e979" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.241818 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hflsg" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.337503 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m"] Oct 10 13:45:41 crc kubenswrapper[4745]: E1010 13:45:41.338070 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="292290b6-c1e2-4b2b-955b-699aa826627d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.338099 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="292290b6-c1e2-4b2b-955b-699aa826627d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 10 13:45:41 crc kubenswrapper[4745]: E1010 13:45:41.338146 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e" containerName="collect-profiles" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.338159 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e" containerName="collect-profiles" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.338480 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="292290b6-c1e2-4b2b-955b-699aa826627d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.338512 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccea7e9b-6e2c-4a68-8a5e-7b5c70a0c68e" containerName="collect-profiles" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.339418 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.342011 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.342155 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.342797 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.342923 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.352572 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m"] Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.385401 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4329eb3b-cd56-417e-9da2-361ab8817091-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m\" (UID: \"4329eb3b-cd56-417e-9da2-361ab8817091\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.385457 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4329eb3b-cd56-417e-9da2-361ab8817091-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m\" (UID: \"4329eb3b-cd56-417e-9da2-361ab8817091\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.385532 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66fss\" (UniqueName: \"kubernetes.io/projected/4329eb3b-cd56-417e-9da2-361ab8817091-kube-api-access-66fss\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m\" (UID: \"4329eb3b-cd56-417e-9da2-361ab8817091\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.487924 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66fss\" (UniqueName: \"kubernetes.io/projected/4329eb3b-cd56-417e-9da2-361ab8817091-kube-api-access-66fss\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m\" (UID: \"4329eb3b-cd56-417e-9da2-361ab8817091\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.488199 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4329eb3b-cd56-417e-9da2-361ab8817091-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m\" (UID: \"4329eb3b-cd56-417e-9da2-361ab8817091\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.488241 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4329eb3b-cd56-417e-9da2-361ab8817091-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m\" (UID: \"4329eb3b-cd56-417e-9da2-361ab8817091\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.493091 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4329eb3b-cd56-417e-9da2-361ab8817091-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m\" (UID: \"4329eb3b-cd56-417e-9da2-361ab8817091\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.493329 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4329eb3b-cd56-417e-9da2-361ab8817091-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m\" (UID: \"4329eb3b-cd56-417e-9da2-361ab8817091\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.517647 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66fss\" (UniqueName: \"kubernetes.io/projected/4329eb3b-cd56-417e-9da2-361ab8817091-kube-api-access-66fss\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m\" (UID: \"4329eb3b-cd56-417e-9da2-361ab8817091\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" Oct 10 13:45:41 crc kubenswrapper[4745]: I1010 13:45:41.659949 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" Oct 10 13:45:42 crc kubenswrapper[4745]: I1010 13:45:42.063314 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m"] Oct 10 13:45:42 crc kubenswrapper[4745]: I1010 13:45:42.252246 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" event={"ID":"4329eb3b-cd56-417e-9da2-361ab8817091","Type":"ContainerStarted","Data":"b06a36d8f2ff25a61ba1479bfbc2c914bead9c1a102579f5c4ab179e495f8e61"} Oct 10 13:45:43 crc kubenswrapper[4745]: I1010 13:45:43.262153 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" event={"ID":"4329eb3b-cd56-417e-9da2-361ab8817091","Type":"ContainerStarted","Data":"d34cf3727190a64de59cba098129411af26cfd86c415ef57c44c480dc9f16c26"} Oct 10 13:45:43 crc kubenswrapper[4745]: I1010 13:45:43.280852 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" podStartSLOduration=1.685464759 podStartE2EDuration="2.280824997s" podCreationTimestamp="2025-10-10 13:45:41 +0000 UTC" firstStartedPulling="2025-10-10 13:45:42.069778256 +0000 UTC m=+1655.967435019" lastFinishedPulling="2025-10-10 13:45:42.665138504 +0000 UTC m=+1656.562795257" observedRunningTime="2025-10-10 13:45:43.277267387 +0000 UTC m=+1657.174924160" watchObservedRunningTime="2025-10-10 13:45:43.280824997 +0000 UTC m=+1657.178481770" Oct 10 13:45:45 crc kubenswrapper[4745]: I1010 13:45:45.744761 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:45:45 crc kubenswrapper[4745]: E1010 13:45:45.745505 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:45:56 crc kubenswrapper[4745]: I1010 13:45:56.759837 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:45:56 crc kubenswrapper[4745]: E1010 13:45:56.760631 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:46:09 crc kubenswrapper[4745]: I1010 13:46:09.745010 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:46:09 crc kubenswrapper[4745]: E1010 13:46:09.745651 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:46:14 crc kubenswrapper[4745]: I1010 13:46:14.403324 4745 scope.go:117] "RemoveContainer" containerID="7852ea056a15f2decd43c0c52a5fb22a6679a34eb1eb98afd32eba762c9edd28" Oct 10 13:46:14 crc kubenswrapper[4745]: I1010 13:46:14.476244 4745 scope.go:117] "RemoveContainer" containerID="c5fbdc3f1f579f71da481b16a039a702ef7b6c33d61cec0f03d77769c1e4cb9e" Oct 10 13:46:14 crc kubenswrapper[4745]: I1010 13:46:14.536904 4745 scope.go:117] "RemoveContainer" containerID="41e66cac9137afc38339955332ce978e26364120893d9f3b693a00b8843750a2" Oct 10 13:46:21 crc kubenswrapper[4745]: I1010 13:46:21.072866 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-ctcrf"] Oct 10 13:46:21 crc kubenswrapper[4745]: I1010 13:46:21.087260 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-ctcrf"] Oct 10 13:46:22 crc kubenswrapper[4745]: I1010 13:46:22.745148 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:46:22 crc kubenswrapper[4745]: E1010 13:46:22.746223 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:46:22 crc kubenswrapper[4745]: I1010 13:46:22.765442 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16ee733c-3513-40bc-97ec-515a9f735a56" path="/var/lib/kubelet/pods/16ee733c-3513-40bc-97ec-515a9f735a56/volumes" Oct 10 13:46:33 crc kubenswrapper[4745]: I1010 13:46:33.746469 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:46:33 crc kubenswrapper[4745]: E1010 13:46:33.747423 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:46:39 crc kubenswrapper[4745]: I1010 13:46:39.831158 4745 generic.go:334] "Generic (PLEG): container finished" podID="4329eb3b-cd56-417e-9da2-361ab8817091" containerID="d34cf3727190a64de59cba098129411af26cfd86c415ef57c44c480dc9f16c26" exitCode=2 Oct 10 13:46:39 crc kubenswrapper[4745]: I1010 13:46:39.831277 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" event={"ID":"4329eb3b-cd56-417e-9da2-361ab8817091","Type":"ContainerDied","Data":"d34cf3727190a64de59cba098129411af26cfd86c415ef57c44c480dc9f16c26"} Oct 10 13:46:41 crc kubenswrapper[4745]: I1010 13:46:41.304935 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" Oct 10 13:46:41 crc kubenswrapper[4745]: I1010 13:46:41.400739 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4329eb3b-cd56-417e-9da2-361ab8817091-inventory\") pod \"4329eb3b-cd56-417e-9da2-361ab8817091\" (UID: \"4329eb3b-cd56-417e-9da2-361ab8817091\") " Oct 10 13:46:41 crc kubenswrapper[4745]: I1010 13:46:41.400802 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66fss\" (UniqueName: \"kubernetes.io/projected/4329eb3b-cd56-417e-9da2-361ab8817091-kube-api-access-66fss\") pod \"4329eb3b-cd56-417e-9da2-361ab8817091\" (UID: \"4329eb3b-cd56-417e-9da2-361ab8817091\") " Oct 10 13:46:41 crc kubenswrapper[4745]: I1010 13:46:41.400914 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4329eb3b-cd56-417e-9da2-361ab8817091-ssh-key\") pod \"4329eb3b-cd56-417e-9da2-361ab8817091\" (UID: \"4329eb3b-cd56-417e-9da2-361ab8817091\") " Oct 10 13:46:41 crc kubenswrapper[4745]: I1010 13:46:41.407445 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4329eb3b-cd56-417e-9da2-361ab8817091-kube-api-access-66fss" (OuterVolumeSpecName: "kube-api-access-66fss") pod "4329eb3b-cd56-417e-9da2-361ab8817091" (UID: "4329eb3b-cd56-417e-9da2-361ab8817091"). InnerVolumeSpecName "kube-api-access-66fss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:46:41 crc kubenswrapper[4745]: I1010 13:46:41.426072 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4329eb3b-cd56-417e-9da2-361ab8817091-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4329eb3b-cd56-417e-9da2-361ab8817091" (UID: "4329eb3b-cd56-417e-9da2-361ab8817091"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:46:41 crc kubenswrapper[4745]: I1010 13:46:41.436915 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4329eb3b-cd56-417e-9da2-361ab8817091-inventory" (OuterVolumeSpecName: "inventory") pod "4329eb3b-cd56-417e-9da2-361ab8817091" (UID: "4329eb3b-cd56-417e-9da2-361ab8817091"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:46:41 crc kubenswrapper[4745]: I1010 13:46:41.502557 4745 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4329eb3b-cd56-417e-9da2-361ab8817091-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 13:46:41 crc kubenswrapper[4745]: I1010 13:46:41.502592 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66fss\" (UniqueName: \"kubernetes.io/projected/4329eb3b-cd56-417e-9da2-361ab8817091-kube-api-access-66fss\") on node \"crc\" DevicePath \"\"" Oct 10 13:46:41 crc kubenswrapper[4745]: I1010 13:46:41.502606 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4329eb3b-cd56-417e-9da2-361ab8817091-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:46:41 crc kubenswrapper[4745]: I1010 13:46:41.853125 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" event={"ID":"4329eb3b-cd56-417e-9da2-361ab8817091","Type":"ContainerDied","Data":"b06a36d8f2ff25a61ba1479bfbc2c914bead9c1a102579f5c4ab179e495f8e61"} Oct 10 13:46:41 crc kubenswrapper[4745]: I1010 13:46:41.853171 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b06a36d8f2ff25a61ba1479bfbc2c914bead9c1a102579f5c4ab179e495f8e61" Oct 10 13:46:41 crc kubenswrapper[4745]: I1010 13:46:41.853172 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m" Oct 10 13:46:45 crc kubenswrapper[4745]: I1010 13:46:45.745192 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:46:45 crc kubenswrapper[4745]: E1010 13:46:45.746104 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.044548 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx"] Oct 10 13:46:49 crc kubenswrapper[4745]: E1010 13:46:49.045713 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4329eb3b-cd56-417e-9da2-361ab8817091" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.045781 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="4329eb3b-cd56-417e-9da2-361ab8817091" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.046245 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="4329eb3b-cd56-417e-9da2-361ab8817091" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.047529 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.050674 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.051953 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.052913 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.053726 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.082528 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx"] Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.145715 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pkxm\" (UniqueName: \"kubernetes.io/projected/0a993d65-ad98-4529-95ad-1663fb206c5a-kube-api-access-4pkxm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx\" (UID: \"0a993d65-ad98-4529-95ad-1663fb206c5a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.145919 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a993d65-ad98-4529-95ad-1663fb206c5a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx\" (UID: \"0a993d65-ad98-4529-95ad-1663fb206c5a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.146176 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a993d65-ad98-4529-95ad-1663fb206c5a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx\" (UID: \"0a993d65-ad98-4529-95ad-1663fb206c5a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.248663 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a993d65-ad98-4529-95ad-1663fb206c5a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx\" (UID: \"0a993d65-ad98-4529-95ad-1663fb206c5a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.248924 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pkxm\" (UniqueName: \"kubernetes.io/projected/0a993d65-ad98-4529-95ad-1663fb206c5a-kube-api-access-4pkxm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx\" (UID: \"0a993d65-ad98-4529-95ad-1663fb206c5a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.249025 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a993d65-ad98-4529-95ad-1663fb206c5a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx\" (UID: \"0a993d65-ad98-4529-95ad-1663fb206c5a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.259244 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a993d65-ad98-4529-95ad-1663fb206c5a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx\" (UID: \"0a993d65-ad98-4529-95ad-1663fb206c5a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.266919 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a993d65-ad98-4529-95ad-1663fb206c5a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx\" (UID: \"0a993d65-ad98-4529-95ad-1663fb206c5a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.271874 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pkxm\" (UniqueName: \"kubernetes.io/projected/0a993d65-ad98-4529-95ad-1663fb206c5a-kube-api-access-4pkxm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx\" (UID: \"0a993d65-ad98-4529-95ad-1663fb206c5a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.381815 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" Oct 10 13:46:49 crc kubenswrapper[4745]: I1010 13:46:49.960576 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx"] Oct 10 13:46:49 crc kubenswrapper[4745]: W1010 13:46:49.968981 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a993d65_ad98_4529_95ad_1663fb206c5a.slice/crio-6f819f6f721f917e782e88fedd336fe20b0c29aabbd2740c25b9dc50ca6ba0a9 WatchSource:0}: Error finding container 6f819f6f721f917e782e88fedd336fe20b0c29aabbd2740c25b9dc50ca6ba0a9: Status 404 returned error can't find the container with id 6f819f6f721f917e782e88fedd336fe20b0c29aabbd2740c25b9dc50ca6ba0a9 Oct 10 13:46:50 crc kubenswrapper[4745]: I1010 13:46:50.959363 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" event={"ID":"0a993d65-ad98-4529-95ad-1663fb206c5a","Type":"ContainerStarted","Data":"3a45da058a240053dbe1ea2bbdea4e04a2dedf56db40e7ca67f2e20c2513fc87"} Oct 10 13:46:50 crc kubenswrapper[4745]: I1010 13:46:50.960708 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" event={"ID":"0a993d65-ad98-4529-95ad-1663fb206c5a","Type":"ContainerStarted","Data":"6f819f6f721f917e782e88fedd336fe20b0c29aabbd2740c25b9dc50ca6ba0a9"} Oct 10 13:46:50 crc kubenswrapper[4745]: I1010 13:46:50.979625 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" podStartSLOduration=1.454081461 podStartE2EDuration="1.97960668s" podCreationTimestamp="2025-10-10 13:46:49 +0000 UTC" firstStartedPulling="2025-10-10 13:46:49.972603919 +0000 UTC m=+1723.870260692" lastFinishedPulling="2025-10-10 13:46:50.498129148 +0000 UTC m=+1724.395785911" observedRunningTime="2025-10-10 13:46:50.976752948 +0000 UTC m=+1724.874409731" watchObservedRunningTime="2025-10-10 13:46:50.97960668 +0000 UTC m=+1724.877263453" Oct 10 13:46:58 crc kubenswrapper[4745]: I1010 13:46:58.749964 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:46:58 crc kubenswrapper[4745]: E1010 13:46:58.750919 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:47:13 crc kubenswrapper[4745]: I1010 13:47:13.745002 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:47:13 crc kubenswrapper[4745]: E1010 13:47:13.745819 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:47:14 crc kubenswrapper[4745]: I1010 13:47:14.630221 4745 scope.go:117] "RemoveContainer" containerID="2e6f2227fb0e20b163ece2be7c2949bb1b87e52d784971061164d05ecc404318" Oct 10 13:47:26 crc kubenswrapper[4745]: I1010 13:47:26.758660 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:47:26 crc kubenswrapper[4745]: E1010 13:47:26.760007 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:47:41 crc kubenswrapper[4745]: I1010 13:47:41.445511 4745 generic.go:334] "Generic (PLEG): container finished" podID="0a993d65-ad98-4529-95ad-1663fb206c5a" containerID="3a45da058a240053dbe1ea2bbdea4e04a2dedf56db40e7ca67f2e20c2513fc87" exitCode=0 Oct 10 13:47:41 crc kubenswrapper[4745]: I1010 13:47:41.445559 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" event={"ID":"0a993d65-ad98-4529-95ad-1663fb206c5a","Type":"ContainerDied","Data":"3a45da058a240053dbe1ea2bbdea4e04a2dedf56db40e7ca67f2e20c2513fc87"} Oct 10 13:47:41 crc kubenswrapper[4745]: I1010 13:47:41.745180 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:47:41 crc kubenswrapper[4745]: E1010 13:47:41.745586 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:47:42 crc kubenswrapper[4745]: I1010 13:47:42.917802 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.089468 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pkxm\" (UniqueName: \"kubernetes.io/projected/0a993d65-ad98-4529-95ad-1663fb206c5a-kube-api-access-4pkxm\") pod \"0a993d65-ad98-4529-95ad-1663fb206c5a\" (UID: \"0a993d65-ad98-4529-95ad-1663fb206c5a\") " Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.089594 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a993d65-ad98-4529-95ad-1663fb206c5a-ssh-key\") pod \"0a993d65-ad98-4529-95ad-1663fb206c5a\" (UID: \"0a993d65-ad98-4529-95ad-1663fb206c5a\") " Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.089663 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a993d65-ad98-4529-95ad-1663fb206c5a-inventory\") pod \"0a993d65-ad98-4529-95ad-1663fb206c5a\" (UID: \"0a993d65-ad98-4529-95ad-1663fb206c5a\") " Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.095940 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a993d65-ad98-4529-95ad-1663fb206c5a-kube-api-access-4pkxm" (OuterVolumeSpecName: "kube-api-access-4pkxm") pod "0a993d65-ad98-4529-95ad-1663fb206c5a" (UID: "0a993d65-ad98-4529-95ad-1663fb206c5a"). InnerVolumeSpecName "kube-api-access-4pkxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.139820 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a993d65-ad98-4529-95ad-1663fb206c5a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0a993d65-ad98-4529-95ad-1663fb206c5a" (UID: "0a993d65-ad98-4529-95ad-1663fb206c5a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.142321 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a993d65-ad98-4529-95ad-1663fb206c5a-inventory" (OuterVolumeSpecName: "inventory") pod "0a993d65-ad98-4529-95ad-1663fb206c5a" (UID: "0a993d65-ad98-4529-95ad-1663fb206c5a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.192134 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pkxm\" (UniqueName: \"kubernetes.io/projected/0a993d65-ad98-4529-95ad-1663fb206c5a-kube-api-access-4pkxm\") on node \"crc\" DevicePath \"\"" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.192179 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a993d65-ad98-4529-95ad-1663fb206c5a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.192192 4745 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a993d65-ad98-4529-95ad-1663fb206c5a-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.468480 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" event={"ID":"0a993d65-ad98-4529-95ad-1663fb206c5a","Type":"ContainerDied","Data":"6f819f6f721f917e782e88fedd336fe20b0c29aabbd2740c25b9dc50ca6ba0a9"} Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.468819 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f819f6f721f917e782e88fedd336fe20b0c29aabbd2740c25b9dc50ca6ba0a9" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.468600 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.561809 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2mslp"] Oct 10 13:47:43 crc kubenswrapper[4745]: E1010 13:47:43.562317 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a993d65-ad98-4529-95ad-1663fb206c5a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.562349 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a993d65-ad98-4529-95ad-1663fb206c5a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.562624 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a993d65-ad98-4529-95ad-1663fb206c5a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.563495 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.565186 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.567146 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.567837 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.568258 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.575792 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2mslp"] Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.701304 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8xk2\" (UniqueName: \"kubernetes.io/projected/0d120415-3123-4b2b-a29d-b1b9c822ed7c-kube-api-access-c8xk2\") pod \"ssh-known-hosts-edpm-deployment-2mslp\" (UID: \"0d120415-3123-4b2b-a29d-b1b9c822ed7c\") " pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.701349 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0d120415-3123-4b2b-a29d-b1b9c822ed7c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2mslp\" (UID: \"0d120415-3123-4b2b-a29d-b1b9c822ed7c\") " pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.701418 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0d120415-3123-4b2b-a29d-b1b9c822ed7c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2mslp\" (UID: \"0d120415-3123-4b2b-a29d-b1b9c822ed7c\") " pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.803190 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8xk2\" (UniqueName: \"kubernetes.io/projected/0d120415-3123-4b2b-a29d-b1b9c822ed7c-kube-api-access-c8xk2\") pod \"ssh-known-hosts-edpm-deployment-2mslp\" (UID: \"0d120415-3123-4b2b-a29d-b1b9c822ed7c\") " pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.803412 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0d120415-3123-4b2b-a29d-b1b9c822ed7c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2mslp\" (UID: \"0d120415-3123-4b2b-a29d-b1b9c822ed7c\") " pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.803571 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0d120415-3123-4b2b-a29d-b1b9c822ed7c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2mslp\" (UID: \"0d120415-3123-4b2b-a29d-b1b9c822ed7c\") " pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.808427 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0d120415-3123-4b2b-a29d-b1b9c822ed7c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2mslp\" (UID: \"0d120415-3123-4b2b-a29d-b1b9c822ed7c\") " pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.809143 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0d120415-3123-4b2b-a29d-b1b9c822ed7c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2mslp\" (UID: \"0d120415-3123-4b2b-a29d-b1b9c822ed7c\") " pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.832595 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8xk2\" (UniqueName: \"kubernetes.io/projected/0d120415-3123-4b2b-a29d-b1b9c822ed7c-kube-api-access-c8xk2\") pod \"ssh-known-hosts-edpm-deployment-2mslp\" (UID: \"0d120415-3123-4b2b-a29d-b1b9c822ed7c\") " pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" Oct 10 13:47:43 crc kubenswrapper[4745]: I1010 13:47:43.887642 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" Oct 10 13:47:44 crc kubenswrapper[4745]: I1010 13:47:44.367913 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2mslp"] Oct 10 13:47:44 crc kubenswrapper[4745]: I1010 13:47:44.508622 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" event={"ID":"0d120415-3123-4b2b-a29d-b1b9c822ed7c","Type":"ContainerStarted","Data":"7dc0bc3e91d0b5e3503bfa5bdc159b5b7cde4120836b58003243c977ee309986"} Oct 10 13:47:45 crc kubenswrapper[4745]: I1010 13:47:45.521876 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" event={"ID":"0d120415-3123-4b2b-a29d-b1b9c822ed7c","Type":"ContainerStarted","Data":"6ed768abd41447ce5f3a4d557e52b2cc7fb095aba7f44e34723e6bd9c7921c81"} Oct 10 13:47:45 crc kubenswrapper[4745]: I1010 13:47:45.547942 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" podStartSLOduration=2.11757118 podStartE2EDuration="2.547921076s" podCreationTimestamp="2025-10-10 13:47:43 +0000 UTC" firstStartedPulling="2025-10-10 13:47:44.373937664 +0000 UTC m=+1778.271594427" lastFinishedPulling="2025-10-10 13:47:44.80428756 +0000 UTC m=+1778.701944323" observedRunningTime="2025-10-10 13:47:45.542250962 +0000 UTC m=+1779.439907765" watchObservedRunningTime="2025-10-10 13:47:45.547921076 +0000 UTC m=+1779.445577839" Oct 10 13:47:52 crc kubenswrapper[4745]: I1010 13:47:52.594209 4745 generic.go:334] "Generic (PLEG): container finished" podID="0d120415-3123-4b2b-a29d-b1b9c822ed7c" containerID="6ed768abd41447ce5f3a4d557e52b2cc7fb095aba7f44e34723e6bd9c7921c81" exitCode=0 Oct 10 13:47:52 crc kubenswrapper[4745]: I1010 13:47:52.594287 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" event={"ID":"0d120415-3123-4b2b-a29d-b1b9c822ed7c","Type":"ContainerDied","Data":"6ed768abd41447ce5f3a4d557e52b2cc7fb095aba7f44e34723e6bd9c7921c81"} Oct 10 13:47:53 crc kubenswrapper[4745]: I1010 13:47:53.745052 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:47:53 crc kubenswrapper[4745]: I1010 13:47:53.995757 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.108310 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0d120415-3123-4b2b-a29d-b1b9c822ed7c-inventory-0\") pod \"0d120415-3123-4b2b-a29d-b1b9c822ed7c\" (UID: \"0d120415-3123-4b2b-a29d-b1b9c822ed7c\") " Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.108407 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0d120415-3123-4b2b-a29d-b1b9c822ed7c-ssh-key-openstack-edpm-ipam\") pod \"0d120415-3123-4b2b-a29d-b1b9c822ed7c\" (UID: \"0d120415-3123-4b2b-a29d-b1b9c822ed7c\") " Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.108435 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8xk2\" (UniqueName: \"kubernetes.io/projected/0d120415-3123-4b2b-a29d-b1b9c822ed7c-kube-api-access-c8xk2\") pod \"0d120415-3123-4b2b-a29d-b1b9c822ed7c\" (UID: \"0d120415-3123-4b2b-a29d-b1b9c822ed7c\") " Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.121897 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d120415-3123-4b2b-a29d-b1b9c822ed7c-kube-api-access-c8xk2" (OuterVolumeSpecName: "kube-api-access-c8xk2") pod "0d120415-3123-4b2b-a29d-b1b9c822ed7c" (UID: "0d120415-3123-4b2b-a29d-b1b9c822ed7c"). InnerVolumeSpecName "kube-api-access-c8xk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.134653 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d120415-3123-4b2b-a29d-b1b9c822ed7c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "0d120415-3123-4b2b-a29d-b1b9c822ed7c" (UID: "0d120415-3123-4b2b-a29d-b1b9c822ed7c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.143407 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d120415-3123-4b2b-a29d-b1b9c822ed7c-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "0d120415-3123-4b2b-a29d-b1b9c822ed7c" (UID: "0d120415-3123-4b2b-a29d-b1b9c822ed7c"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.211031 4745 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0d120415-3123-4b2b-a29d-b1b9c822ed7c-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.211060 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0d120415-3123-4b2b-a29d-b1b9c822ed7c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.211070 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8xk2\" (UniqueName: \"kubernetes.io/projected/0d120415-3123-4b2b-a29d-b1b9c822ed7c-kube-api-access-c8xk2\") on node \"crc\" DevicePath \"\"" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.616338 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"ca4a6f66c74821d4fbc4213650a195182b8af7f523a58adff73b04160e45bbe7"} Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.620256 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" event={"ID":"0d120415-3123-4b2b-a29d-b1b9c822ed7c","Type":"ContainerDied","Data":"7dc0bc3e91d0b5e3503bfa5bdc159b5b7cde4120836b58003243c977ee309986"} Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.620298 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dc0bc3e91d0b5e3503bfa5bdc159b5b7cde4120836b58003243c977ee309986" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.620313 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2mslp" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.766914 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp"] Oct 10 13:47:54 crc kubenswrapper[4745]: E1010 13:47:54.767260 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d120415-3123-4b2b-a29d-b1b9c822ed7c" containerName="ssh-known-hosts-edpm-deployment" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.767274 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d120415-3123-4b2b-a29d-b1b9c822ed7c" containerName="ssh-known-hosts-edpm-deployment" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.767486 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d120415-3123-4b2b-a29d-b1b9c822ed7c" containerName="ssh-known-hosts-edpm-deployment" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.768123 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.770444 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.771050 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.771232 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.771453 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.781094 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp"] Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.821169 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29cef4fe-f039-411b-a979-91362f56cfd7-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vr2vp\" (UID: \"29cef4fe-f039-411b-a979-91362f56cfd7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.821456 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhl75\" (UniqueName: \"kubernetes.io/projected/29cef4fe-f039-411b-a979-91362f56cfd7-kube-api-access-lhl75\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vr2vp\" (UID: \"29cef4fe-f039-411b-a979-91362f56cfd7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.821683 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29cef4fe-f039-411b-a979-91362f56cfd7-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vr2vp\" (UID: \"29cef4fe-f039-411b-a979-91362f56cfd7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.924070 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29cef4fe-f039-411b-a979-91362f56cfd7-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vr2vp\" (UID: \"29cef4fe-f039-411b-a979-91362f56cfd7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.924571 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29cef4fe-f039-411b-a979-91362f56cfd7-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vr2vp\" (UID: \"29cef4fe-f039-411b-a979-91362f56cfd7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.925389 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhl75\" (UniqueName: \"kubernetes.io/projected/29cef4fe-f039-411b-a979-91362f56cfd7-kube-api-access-lhl75\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vr2vp\" (UID: \"29cef4fe-f039-411b-a979-91362f56cfd7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.931648 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29cef4fe-f039-411b-a979-91362f56cfd7-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vr2vp\" (UID: \"29cef4fe-f039-411b-a979-91362f56cfd7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.931651 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29cef4fe-f039-411b-a979-91362f56cfd7-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vr2vp\" (UID: \"29cef4fe-f039-411b-a979-91362f56cfd7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" Oct 10 13:47:54 crc kubenswrapper[4745]: I1010 13:47:54.944417 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhl75\" (UniqueName: \"kubernetes.io/projected/29cef4fe-f039-411b-a979-91362f56cfd7-kube-api-access-lhl75\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vr2vp\" (UID: \"29cef4fe-f039-411b-a979-91362f56cfd7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" Oct 10 13:47:55 crc kubenswrapper[4745]: I1010 13:47:55.089897 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" Oct 10 13:47:55 crc kubenswrapper[4745]: I1010 13:47:55.606036 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp"] Oct 10 13:47:55 crc kubenswrapper[4745]: W1010 13:47:55.615528 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29cef4fe_f039_411b_a979_91362f56cfd7.slice/crio-372265d54ce140d08cb442a69d55b679069b3d3276f11015dcbf9848c1bc23af WatchSource:0}: Error finding container 372265d54ce140d08cb442a69d55b679069b3d3276f11015dcbf9848c1bc23af: Status 404 returned error can't find the container with id 372265d54ce140d08cb442a69d55b679069b3d3276f11015dcbf9848c1bc23af Oct 10 13:47:55 crc kubenswrapper[4745]: I1010 13:47:55.636120 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" event={"ID":"29cef4fe-f039-411b-a979-91362f56cfd7","Type":"ContainerStarted","Data":"372265d54ce140d08cb442a69d55b679069b3d3276f11015dcbf9848c1bc23af"} Oct 10 13:47:56 crc kubenswrapper[4745]: I1010 13:47:56.646437 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" event={"ID":"29cef4fe-f039-411b-a979-91362f56cfd7","Type":"ContainerStarted","Data":"af1cfd3132dbe9169664d3614d53adfb7bce7c5b6d3fe5fae357219843fdd763"} Oct 10 13:47:56 crc kubenswrapper[4745]: I1010 13:47:56.673996 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" podStartSLOduration=2.167144265 podStartE2EDuration="2.673969309s" podCreationTimestamp="2025-10-10 13:47:54 +0000 UTC" firstStartedPulling="2025-10-10 13:47:55.625042336 +0000 UTC m=+1789.522699119" lastFinishedPulling="2025-10-10 13:47:56.13186739 +0000 UTC m=+1790.029524163" observedRunningTime="2025-10-10 13:47:56.667534226 +0000 UTC m=+1790.565191069" watchObservedRunningTime="2025-10-10 13:47:56.673969309 +0000 UTC m=+1790.571626092" Oct 10 13:48:05 crc kubenswrapper[4745]: I1010 13:48:05.744518 4745 generic.go:334] "Generic (PLEG): container finished" podID="29cef4fe-f039-411b-a979-91362f56cfd7" containerID="af1cfd3132dbe9169664d3614d53adfb7bce7c5b6d3fe5fae357219843fdd763" exitCode=0 Oct 10 13:48:05 crc kubenswrapper[4745]: I1010 13:48:05.744606 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" event={"ID":"29cef4fe-f039-411b-a979-91362f56cfd7","Type":"ContainerDied","Data":"af1cfd3132dbe9169664d3614d53adfb7bce7c5b6d3fe5fae357219843fdd763"} Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.154124 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.279185 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29cef4fe-f039-411b-a979-91362f56cfd7-ssh-key\") pod \"29cef4fe-f039-411b-a979-91362f56cfd7\" (UID: \"29cef4fe-f039-411b-a979-91362f56cfd7\") " Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.279318 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29cef4fe-f039-411b-a979-91362f56cfd7-inventory\") pod \"29cef4fe-f039-411b-a979-91362f56cfd7\" (UID: \"29cef4fe-f039-411b-a979-91362f56cfd7\") " Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.279367 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhl75\" (UniqueName: \"kubernetes.io/projected/29cef4fe-f039-411b-a979-91362f56cfd7-kube-api-access-lhl75\") pod \"29cef4fe-f039-411b-a979-91362f56cfd7\" (UID: \"29cef4fe-f039-411b-a979-91362f56cfd7\") " Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.285593 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29cef4fe-f039-411b-a979-91362f56cfd7-kube-api-access-lhl75" (OuterVolumeSpecName: "kube-api-access-lhl75") pod "29cef4fe-f039-411b-a979-91362f56cfd7" (UID: "29cef4fe-f039-411b-a979-91362f56cfd7"). InnerVolumeSpecName "kube-api-access-lhl75". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.310664 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29cef4fe-f039-411b-a979-91362f56cfd7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "29cef4fe-f039-411b-a979-91362f56cfd7" (UID: "29cef4fe-f039-411b-a979-91362f56cfd7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.315410 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29cef4fe-f039-411b-a979-91362f56cfd7-inventory" (OuterVolumeSpecName: "inventory") pod "29cef4fe-f039-411b-a979-91362f56cfd7" (UID: "29cef4fe-f039-411b-a979-91362f56cfd7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.381498 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29cef4fe-f039-411b-a979-91362f56cfd7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.381538 4745 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29cef4fe-f039-411b-a979-91362f56cfd7-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.381552 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhl75\" (UniqueName: \"kubernetes.io/projected/29cef4fe-f039-411b-a979-91362f56cfd7-kube-api-access-lhl75\") on node \"crc\" DevicePath \"\"" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.765159 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" event={"ID":"29cef4fe-f039-411b-a979-91362f56cfd7","Type":"ContainerDied","Data":"372265d54ce140d08cb442a69d55b679069b3d3276f11015dcbf9848c1bc23af"} Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.765416 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="372265d54ce140d08cb442a69d55b679069b3d3276f11015dcbf9848c1bc23af" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.765241 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vr2vp" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.850044 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n"] Oct 10 13:48:07 crc kubenswrapper[4745]: E1010 13:48:07.850487 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29cef4fe-f039-411b-a979-91362f56cfd7" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.850511 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="29cef4fe-f039-411b-a979-91362f56cfd7" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.852533 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="29cef4fe-f039-411b-a979-91362f56cfd7" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.853295 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.856710 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.856955 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.859099 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.859543 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.913885 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n"] Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.991724 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n\" (UID: \"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.991853 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n\" (UID: \"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" Oct 10 13:48:07 crc kubenswrapper[4745]: I1010 13:48:07.991890 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prvln\" (UniqueName: \"kubernetes.io/projected/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-kube-api-access-prvln\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n\" (UID: \"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" Oct 10 13:48:08 crc kubenswrapper[4745]: I1010 13:48:08.093975 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prvln\" (UniqueName: \"kubernetes.io/projected/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-kube-api-access-prvln\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n\" (UID: \"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" Oct 10 13:48:08 crc kubenswrapper[4745]: I1010 13:48:08.094110 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n\" (UID: \"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" Oct 10 13:48:08 crc kubenswrapper[4745]: I1010 13:48:08.094217 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n\" (UID: \"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" Oct 10 13:48:08 crc kubenswrapper[4745]: I1010 13:48:08.100030 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n\" (UID: \"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" Oct 10 13:48:08 crc kubenswrapper[4745]: I1010 13:48:08.100266 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n\" (UID: \"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" Oct 10 13:48:08 crc kubenswrapper[4745]: I1010 13:48:08.111397 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prvln\" (UniqueName: \"kubernetes.io/projected/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-kube-api-access-prvln\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n\" (UID: \"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" Oct 10 13:48:08 crc kubenswrapper[4745]: I1010 13:48:08.172527 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" Oct 10 13:48:08 crc kubenswrapper[4745]: I1010 13:48:08.673361 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n"] Oct 10 13:48:08 crc kubenswrapper[4745]: W1010 13:48:08.681542 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb812bcc0_a0b6_4ef6_afd2_d0b8234f1ab4.slice/crio-a548d611b0e1df7e73618bb7e33f8c8ff13151f51f0614a00043aaa631364bfd WatchSource:0}: Error finding container a548d611b0e1df7e73618bb7e33f8c8ff13151f51f0614a00043aaa631364bfd: Status 404 returned error can't find the container with id a548d611b0e1df7e73618bb7e33f8c8ff13151f51f0614a00043aaa631364bfd Oct 10 13:48:08 crc kubenswrapper[4745]: I1010 13:48:08.774458 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" event={"ID":"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4","Type":"ContainerStarted","Data":"a548d611b0e1df7e73618bb7e33f8c8ff13151f51f0614a00043aaa631364bfd"} Oct 10 13:48:09 crc kubenswrapper[4745]: I1010 13:48:09.786942 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" event={"ID":"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4","Type":"ContainerStarted","Data":"9dee6a809c8c709dbacad4712e316ab9b23e179a48c0fc02a3ff97b5eaa8d632"} Oct 10 13:48:09 crc kubenswrapper[4745]: I1010 13:48:09.806105 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" podStartSLOduration=2.3122870779999998 podStartE2EDuration="2.806084112s" podCreationTimestamp="2025-10-10 13:48:07 +0000 UTC" firstStartedPulling="2025-10-10 13:48:08.685194735 +0000 UTC m=+1802.582851518" lastFinishedPulling="2025-10-10 13:48:09.178991749 +0000 UTC m=+1803.076648552" observedRunningTime="2025-10-10 13:48:09.800367587 +0000 UTC m=+1803.698024360" watchObservedRunningTime="2025-10-10 13:48:09.806084112 +0000 UTC m=+1803.703740865" Oct 10 13:48:19 crc kubenswrapper[4745]: I1010 13:48:19.896349 4745 generic.go:334] "Generic (PLEG): container finished" podID="b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4" containerID="9dee6a809c8c709dbacad4712e316ab9b23e179a48c0fc02a3ff97b5eaa8d632" exitCode=0 Oct 10 13:48:19 crc kubenswrapper[4745]: I1010 13:48:19.897219 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" event={"ID":"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4","Type":"ContainerDied","Data":"9dee6a809c8c709dbacad4712e316ab9b23e179a48c0fc02a3ff97b5eaa8d632"} Oct 10 13:48:21 crc kubenswrapper[4745]: I1010 13:48:21.470540 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" Oct 10 13:48:21 crc kubenswrapper[4745]: I1010 13:48:21.591670 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-ssh-key\") pod \"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4\" (UID: \"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4\") " Oct 10 13:48:21 crc kubenswrapper[4745]: I1010 13:48:21.591744 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prvln\" (UniqueName: \"kubernetes.io/projected/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-kube-api-access-prvln\") pod \"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4\" (UID: \"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4\") " Oct 10 13:48:21 crc kubenswrapper[4745]: I1010 13:48:21.591830 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-inventory\") pod \"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4\" (UID: \"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4\") " Oct 10 13:48:21 crc kubenswrapper[4745]: I1010 13:48:21.596885 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-kube-api-access-prvln" (OuterVolumeSpecName: "kube-api-access-prvln") pod "b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4" (UID: "b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4"). InnerVolumeSpecName "kube-api-access-prvln". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:48:21 crc kubenswrapper[4745]: I1010 13:48:21.628198 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-inventory" (OuterVolumeSpecName: "inventory") pod "b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4" (UID: "b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:48:21 crc kubenswrapper[4745]: I1010 13:48:21.641203 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4" (UID: "b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:48:21 crc kubenswrapper[4745]: I1010 13:48:21.693885 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:48:21 crc kubenswrapper[4745]: I1010 13:48:21.693913 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prvln\" (UniqueName: \"kubernetes.io/projected/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-kube-api-access-prvln\") on node \"crc\" DevicePath \"\"" Oct 10 13:48:21 crc kubenswrapper[4745]: I1010 13:48:21.693923 4745 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 13:48:21 crc kubenswrapper[4745]: I1010 13:48:21.918246 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" event={"ID":"b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4","Type":"ContainerDied","Data":"a548d611b0e1df7e73618bb7e33f8c8ff13151f51f0614a00043aaa631364bfd"} Oct 10 13:48:21 crc kubenswrapper[4745]: I1010 13:48:21.918284 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a548d611b0e1df7e73618bb7e33f8c8ff13151f51f0614a00043aaa631364bfd" Oct 10 13:48:21 crc kubenswrapper[4745]: I1010 13:48:21.918339 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.004745 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx"] Oct 10 13:48:22 crc kubenswrapper[4745]: E1010 13:48:22.005154 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.005176 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.005410 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.006148 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.008288 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.008769 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.008955 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.009051 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.009218 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.009234 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.009309 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.009462 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.037367 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx"] Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.104082 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.104144 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.104203 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.104227 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.104280 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.104344 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.104372 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.104396 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.104434 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.104505 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.104560 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfbbf\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-kube-api-access-mfbbf\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.104584 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.104621 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.104649 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.206087 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.206138 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.206159 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.206190 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.206237 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.206274 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfbbf\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-kube-api-access-mfbbf\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.206292 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.206318 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.206341 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.206361 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.206377 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.206406 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.206425 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.206456 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.218422 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.218474 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.223637 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.223685 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.223851 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.223952 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.224148 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.224485 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.225086 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.227549 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.229285 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.230456 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.235665 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.253676 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfbbf\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-kube-api-access-mfbbf\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rngcx\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.322762 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.836764 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx"] Oct 10 13:48:22 crc kubenswrapper[4745]: I1010 13:48:22.930822 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" event={"ID":"8360e97b-9df5-4790-8796-06c6280f323d","Type":"ContainerStarted","Data":"4ed2f4b0326f516ad3ca1368d00debde1158e796e4a07aa4908c237582b08867"} Oct 10 13:48:23 crc kubenswrapper[4745]: I1010 13:48:23.943338 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" event={"ID":"8360e97b-9df5-4790-8796-06c6280f323d","Type":"ContainerStarted","Data":"136442c8031b4e4c4849b081bb34e5cf8c31e9d66e88ec864f82cd3a27b3f5a7"} Oct 10 13:48:23 crc kubenswrapper[4745]: I1010 13:48:23.976512 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" podStartSLOduration=2.470900583 podStartE2EDuration="2.976465635s" podCreationTimestamp="2025-10-10 13:48:21 +0000 UTC" firstStartedPulling="2025-10-10 13:48:22.842405436 +0000 UTC m=+1816.740062199" lastFinishedPulling="2025-10-10 13:48:23.347970468 +0000 UTC m=+1817.245627251" observedRunningTime="2025-10-10 13:48:23.962225324 +0000 UTC m=+1817.859882097" watchObservedRunningTime="2025-10-10 13:48:23.976465635 +0000 UTC m=+1817.874122408" Oct 10 13:49:04 crc kubenswrapper[4745]: I1010 13:49:04.325798 4745 generic.go:334] "Generic (PLEG): container finished" podID="8360e97b-9df5-4790-8796-06c6280f323d" containerID="136442c8031b4e4c4849b081bb34e5cf8c31e9d66e88ec864f82cd3a27b3f5a7" exitCode=0 Oct 10 13:49:04 crc kubenswrapper[4745]: I1010 13:49:04.325888 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" event={"ID":"8360e97b-9df5-4790-8796-06c6280f323d","Type":"ContainerDied","Data":"136442c8031b4e4c4849b081bb34e5cf8c31e9d66e88ec864f82cd3a27b3f5a7"} Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.737467 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.794236 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-bootstrap-combined-ca-bundle\") pod \"8360e97b-9df5-4790-8796-06c6280f323d\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.794557 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfbbf\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-kube-api-access-mfbbf\") pod \"8360e97b-9df5-4790-8796-06c6280f323d\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.794582 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-inventory\") pod \"8360e97b-9df5-4790-8796-06c6280f323d\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.794652 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-libvirt-combined-ca-bundle\") pod \"8360e97b-9df5-4790-8796-06c6280f323d\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.794677 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"8360e97b-9df5-4790-8796-06c6280f323d\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.794724 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"8360e97b-9df5-4790-8796-06c6280f323d\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.794891 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-repo-setup-combined-ca-bundle\") pod \"8360e97b-9df5-4790-8796-06c6280f323d\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.794931 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-neutron-metadata-combined-ca-bundle\") pod \"8360e97b-9df5-4790-8796-06c6280f323d\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.794957 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-telemetry-combined-ca-bundle\") pod \"8360e97b-9df5-4790-8796-06c6280f323d\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.794995 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"8360e97b-9df5-4790-8796-06c6280f323d\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.795029 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-ovn-combined-ca-bundle\") pod \"8360e97b-9df5-4790-8796-06c6280f323d\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.795095 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-ssh-key\") pod \"8360e97b-9df5-4790-8796-06c6280f323d\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.795131 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-nova-combined-ca-bundle\") pod \"8360e97b-9df5-4790-8796-06c6280f323d\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.795161 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"8360e97b-9df5-4790-8796-06c6280f323d\" (UID: \"8360e97b-9df5-4790-8796-06c6280f323d\") " Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.802782 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "8360e97b-9df5-4790-8796-06c6280f323d" (UID: "8360e97b-9df5-4790-8796-06c6280f323d"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.805774 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "8360e97b-9df5-4790-8796-06c6280f323d" (UID: "8360e97b-9df5-4790-8796-06c6280f323d"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.805871 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "8360e97b-9df5-4790-8796-06c6280f323d" (UID: "8360e97b-9df5-4790-8796-06c6280f323d"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.806267 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "8360e97b-9df5-4790-8796-06c6280f323d" (UID: "8360e97b-9df5-4790-8796-06c6280f323d"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.806288 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "8360e97b-9df5-4790-8796-06c6280f323d" (UID: "8360e97b-9df5-4790-8796-06c6280f323d"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.806324 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "8360e97b-9df5-4790-8796-06c6280f323d" (UID: "8360e97b-9df5-4790-8796-06c6280f323d"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.806349 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-kube-api-access-mfbbf" (OuterVolumeSpecName: "kube-api-access-mfbbf") pod "8360e97b-9df5-4790-8796-06c6280f323d" (UID: "8360e97b-9df5-4790-8796-06c6280f323d"). InnerVolumeSpecName "kube-api-access-mfbbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.806383 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "8360e97b-9df5-4790-8796-06c6280f323d" (UID: "8360e97b-9df5-4790-8796-06c6280f323d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.809708 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "8360e97b-9df5-4790-8796-06c6280f323d" (UID: "8360e97b-9df5-4790-8796-06c6280f323d"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.812175 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "8360e97b-9df5-4790-8796-06c6280f323d" (UID: "8360e97b-9df5-4790-8796-06c6280f323d"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.812540 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "8360e97b-9df5-4790-8796-06c6280f323d" (UID: "8360e97b-9df5-4790-8796-06c6280f323d"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.816850 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "8360e97b-9df5-4790-8796-06c6280f323d" (UID: "8360e97b-9df5-4790-8796-06c6280f323d"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.831330 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-inventory" (OuterVolumeSpecName: "inventory") pod "8360e97b-9df5-4790-8796-06c6280f323d" (UID: "8360e97b-9df5-4790-8796-06c6280f323d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.836015 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8360e97b-9df5-4790-8796-06c6280f323d" (UID: "8360e97b-9df5-4790-8796-06c6280f323d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.897047 4745 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.897089 4745 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.897108 4745 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.897130 4745 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.897195 4745 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.897249 4745 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.897264 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.897278 4745 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.897293 4745 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.897308 4745 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.897320 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfbbf\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-kube-api-access-mfbbf\") on node \"crc\" DevicePath \"\"" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.897331 4745 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.897341 4745 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8360e97b-9df5-4790-8796-06c6280f323d-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:49:05 crc kubenswrapper[4745]: I1010 13:49:05.897354 4745 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8360e97b-9df5-4790-8796-06c6280f323d-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.349854 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" event={"ID":"8360e97b-9df5-4790-8796-06c6280f323d","Type":"ContainerDied","Data":"4ed2f4b0326f516ad3ca1368d00debde1158e796e4a07aa4908c237582b08867"} Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.349902 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ed2f4b0326f516ad3ca1368d00debde1158e796e4a07aa4908c237582b08867" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.349963 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rngcx" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.495715 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h"] Oct 10 13:49:06 crc kubenswrapper[4745]: E1010 13:49:06.496590 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8360e97b-9df5-4790-8796-06c6280f323d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.496612 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="8360e97b-9df5-4790-8796-06c6280f323d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.497248 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="8360e97b-9df5-4790-8796-06c6280f323d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.498373 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.501532 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.501702 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.501821 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.501935 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.505476 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.532657 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h"] Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.612247 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ca7c907d-5166-4e04-b485-04d0b903d6cd-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5fz6h\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.612323 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5fz6h\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.612353 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5fz6h\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.612392 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d2x2\" (UniqueName: \"kubernetes.io/projected/ca7c907d-5166-4e04-b485-04d0b903d6cd-kube-api-access-9d2x2\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5fz6h\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.612422 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5fz6h\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.714796 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ca7c907d-5166-4e04-b485-04d0b903d6cd-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5fz6h\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.714953 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5fz6h\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.715004 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5fz6h\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.715062 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d2x2\" (UniqueName: \"kubernetes.io/projected/ca7c907d-5166-4e04-b485-04d0b903d6cd-kube-api-access-9d2x2\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5fz6h\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.715112 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5fz6h\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.718349 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.718363 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.718460 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.721642 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5fz6h\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.726820 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ca7c907d-5166-4e04-b485-04d0b903d6cd-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5fz6h\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.729864 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5fz6h\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.730205 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5fz6h\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.736565 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d2x2\" (UniqueName: \"kubernetes.io/projected/ca7c907d-5166-4e04-b485-04d0b903d6cd-kube-api-access-9d2x2\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5fz6h\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.842173 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:49:06 crc kubenswrapper[4745]: I1010 13:49:06.850885 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:49:07 crc kubenswrapper[4745]: I1010 13:49:07.398897 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h"] Oct 10 13:49:07 crc kubenswrapper[4745]: I1010 13:49:07.408950 4745 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 13:49:07 crc kubenswrapper[4745]: I1010 13:49:07.919077 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:49:08 crc kubenswrapper[4745]: I1010 13:49:08.381926 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" event={"ID":"ca7c907d-5166-4e04-b485-04d0b903d6cd","Type":"ContainerStarted","Data":"16719d72bf06c77e694eaffe5ccf9cfa179b44220c1f7206c6f666609c5184a8"} Oct 10 13:49:08 crc kubenswrapper[4745]: I1010 13:49:08.381970 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" event={"ID":"ca7c907d-5166-4e04-b485-04d0b903d6cd","Type":"ContainerStarted","Data":"975bcb209d522877fd8446040858344290c41d846686ace47b174d021ae2e08d"} Oct 10 13:49:08 crc kubenswrapper[4745]: I1010 13:49:08.405966 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" podStartSLOduration=1.897676876 podStartE2EDuration="2.405946347s" podCreationTimestamp="2025-10-10 13:49:06 +0000 UTC" firstStartedPulling="2025-10-10 13:49:07.408685514 +0000 UTC m=+1861.306342277" lastFinishedPulling="2025-10-10 13:49:07.916954985 +0000 UTC m=+1861.814611748" observedRunningTime="2025-10-10 13:49:08.399054743 +0000 UTC m=+1862.296711516" watchObservedRunningTime="2025-10-10 13:49:08.405946347 +0000 UTC m=+1862.303603110" Oct 10 13:50:16 crc kubenswrapper[4745]: I1010 13:50:16.066980 4745 generic.go:334] "Generic (PLEG): container finished" podID="ca7c907d-5166-4e04-b485-04d0b903d6cd" containerID="16719d72bf06c77e694eaffe5ccf9cfa179b44220c1f7206c6f666609c5184a8" exitCode=0 Oct 10 13:50:16 crc kubenswrapper[4745]: I1010 13:50:16.067073 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" event={"ID":"ca7c907d-5166-4e04-b485-04d0b903d6cd","Type":"ContainerDied","Data":"16719d72bf06c77e694eaffe5ccf9cfa179b44220c1f7206c6f666609c5184a8"} Oct 10 13:50:16 crc kubenswrapper[4745]: I1010 13:50:16.186439 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:50:16 crc kubenswrapper[4745]: I1010 13:50:16.186850 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:50:17 crc kubenswrapper[4745]: I1010 13:50:17.553939 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:50:17 crc kubenswrapper[4745]: I1010 13:50:17.690951 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-ssh-key\") pod \"ca7c907d-5166-4e04-b485-04d0b903d6cd\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " Oct 10 13:50:17 crc kubenswrapper[4745]: I1010 13:50:17.691765 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ca7c907d-5166-4e04-b485-04d0b903d6cd-ovncontroller-config-0\") pod \"ca7c907d-5166-4e04-b485-04d0b903d6cd\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " Oct 10 13:50:17 crc kubenswrapper[4745]: I1010 13:50:17.691808 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-ovn-combined-ca-bundle\") pod \"ca7c907d-5166-4e04-b485-04d0b903d6cd\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " Oct 10 13:50:17 crc kubenswrapper[4745]: I1010 13:50:17.691915 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9d2x2\" (UniqueName: \"kubernetes.io/projected/ca7c907d-5166-4e04-b485-04d0b903d6cd-kube-api-access-9d2x2\") pod \"ca7c907d-5166-4e04-b485-04d0b903d6cd\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " Oct 10 13:50:17 crc kubenswrapper[4745]: I1010 13:50:17.692040 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-inventory\") pod \"ca7c907d-5166-4e04-b485-04d0b903d6cd\" (UID: \"ca7c907d-5166-4e04-b485-04d0b903d6cd\") " Oct 10 13:50:17 crc kubenswrapper[4745]: I1010 13:50:17.695972 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca7c907d-5166-4e04-b485-04d0b903d6cd-kube-api-access-9d2x2" (OuterVolumeSpecName: "kube-api-access-9d2x2") pod "ca7c907d-5166-4e04-b485-04d0b903d6cd" (UID: "ca7c907d-5166-4e04-b485-04d0b903d6cd"). InnerVolumeSpecName "kube-api-access-9d2x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:50:17 crc kubenswrapper[4745]: I1010 13:50:17.698221 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "ca7c907d-5166-4e04-b485-04d0b903d6cd" (UID: "ca7c907d-5166-4e04-b485-04d0b903d6cd"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:50:17 crc kubenswrapper[4745]: I1010 13:50:17.716905 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-inventory" (OuterVolumeSpecName: "inventory") pod "ca7c907d-5166-4e04-b485-04d0b903d6cd" (UID: "ca7c907d-5166-4e04-b485-04d0b903d6cd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:50:17 crc kubenswrapper[4745]: I1010 13:50:17.719341 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca7c907d-5166-4e04-b485-04d0b903d6cd-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "ca7c907d-5166-4e04-b485-04d0b903d6cd" (UID: "ca7c907d-5166-4e04-b485-04d0b903d6cd"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:50:17 crc kubenswrapper[4745]: I1010 13:50:17.727046 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ca7c907d-5166-4e04-b485-04d0b903d6cd" (UID: "ca7c907d-5166-4e04-b485-04d0b903d6cd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:50:17 crc kubenswrapper[4745]: I1010 13:50:17.793744 4745 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ca7c907d-5166-4e04-b485-04d0b903d6cd-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:50:17 crc kubenswrapper[4745]: I1010 13:50:17.793780 4745 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:50:17 crc kubenswrapper[4745]: I1010 13:50:17.793799 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9d2x2\" (UniqueName: \"kubernetes.io/projected/ca7c907d-5166-4e04-b485-04d0b903d6cd-kube-api-access-9d2x2\") on node \"crc\" DevicePath \"\"" Oct 10 13:50:17 crc kubenswrapper[4745]: I1010 13:50:17.793808 4745 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 13:50:17 crc kubenswrapper[4745]: I1010 13:50:17.794686 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca7c907d-5166-4e04-b485-04d0b903d6cd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.101455 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" event={"ID":"ca7c907d-5166-4e04-b485-04d0b903d6cd","Type":"ContainerDied","Data":"975bcb209d522877fd8446040858344290c41d846686ace47b174d021ae2e08d"} Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.101839 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="975bcb209d522877fd8446040858344290c41d846686ace47b174d021ae2e08d" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.101550 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5fz6h" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.271355 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2"] Oct 10 13:50:18 crc kubenswrapper[4745]: E1010 13:50:18.271750 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca7c907d-5166-4e04-b485-04d0b903d6cd" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.271764 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca7c907d-5166-4e04-b485-04d0b903d6cd" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.271937 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca7c907d-5166-4e04-b485-04d0b903d6cd" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.272534 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.277092 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.277166 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.277202 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.277418 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.277497 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.277773 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.285872 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2"] Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.407949 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.407982 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.408029 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.408052 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.408092 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.408167 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfzh2\" (UniqueName: \"kubernetes.io/projected/4674795d-7ada-40eb-93d6-524c7af53ee3-kube-api-access-sfzh2\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.510285 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfzh2\" (UniqueName: \"kubernetes.io/projected/4674795d-7ada-40eb-93d6-524c7af53ee3-kube-api-access-sfzh2\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.510380 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.510421 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.510463 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.510500 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.510537 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.515180 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.516821 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.517611 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.520007 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.525532 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.528855 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfzh2\" (UniqueName: \"kubernetes.io/projected/4674795d-7ada-40eb-93d6-524c7af53ee3-kube-api-access-sfzh2\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:18 crc kubenswrapper[4745]: I1010 13:50:18.600782 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:50:19 crc kubenswrapper[4745]: I1010 13:50:19.141617 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2"] Oct 10 13:50:20 crc kubenswrapper[4745]: I1010 13:50:20.137352 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" event={"ID":"4674795d-7ada-40eb-93d6-524c7af53ee3","Type":"ContainerStarted","Data":"da19c6273e7e199f36337918b0383102c643b31be3b5a8612b3c2668c8aa76a9"} Oct 10 13:50:20 crc kubenswrapper[4745]: I1010 13:50:20.137607 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" event={"ID":"4674795d-7ada-40eb-93d6-524c7af53ee3","Type":"ContainerStarted","Data":"0bcf1fd13a4a67b92dac799531a1b5d24ffa3241adf05f255983bf3eec79a037"} Oct 10 13:50:20 crc kubenswrapper[4745]: I1010 13:50:20.166306 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" podStartSLOduration=1.589711976 podStartE2EDuration="2.166289441s" podCreationTimestamp="2025-10-10 13:50:18 +0000 UTC" firstStartedPulling="2025-10-10 13:50:19.141453302 +0000 UTC m=+1933.039110065" lastFinishedPulling="2025-10-10 13:50:19.718030727 +0000 UTC m=+1933.615687530" observedRunningTime="2025-10-10 13:50:20.162703621 +0000 UTC m=+1934.060360424" watchObservedRunningTime="2025-10-10 13:50:20.166289441 +0000 UTC m=+1934.063946204" Oct 10 13:50:46 crc kubenswrapper[4745]: I1010 13:50:46.186585 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:50:46 crc kubenswrapper[4745]: I1010 13:50:46.187131 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:51:11 crc kubenswrapper[4745]: I1010 13:51:11.653154 4745 generic.go:334] "Generic (PLEG): container finished" podID="4674795d-7ada-40eb-93d6-524c7af53ee3" containerID="da19c6273e7e199f36337918b0383102c643b31be3b5a8612b3c2668c8aa76a9" exitCode=0 Oct 10 13:51:11 crc kubenswrapper[4745]: I1010 13:51:11.653218 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" event={"ID":"4674795d-7ada-40eb-93d6-524c7af53ee3","Type":"ContainerDied","Data":"da19c6273e7e199f36337918b0383102c643b31be3b5a8612b3c2668c8aa76a9"} Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.086410 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.284114 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"4674795d-7ada-40eb-93d6-524c7af53ee3\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.284237 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-nova-metadata-neutron-config-0\") pod \"4674795d-7ada-40eb-93d6-524c7af53ee3\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.284402 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-ssh-key\") pod \"4674795d-7ada-40eb-93d6-524c7af53ee3\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.284511 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfzh2\" (UniqueName: \"kubernetes.io/projected/4674795d-7ada-40eb-93d6-524c7af53ee3-kube-api-access-sfzh2\") pod \"4674795d-7ada-40eb-93d6-524c7af53ee3\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.284548 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-neutron-metadata-combined-ca-bundle\") pod \"4674795d-7ada-40eb-93d6-524c7af53ee3\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.284770 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-inventory\") pod \"4674795d-7ada-40eb-93d6-524c7af53ee3\" (UID: \"4674795d-7ada-40eb-93d6-524c7af53ee3\") " Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.291252 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4674795d-7ada-40eb-93d6-524c7af53ee3-kube-api-access-sfzh2" (OuterVolumeSpecName: "kube-api-access-sfzh2") pod "4674795d-7ada-40eb-93d6-524c7af53ee3" (UID: "4674795d-7ada-40eb-93d6-524c7af53ee3"). InnerVolumeSpecName "kube-api-access-sfzh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.292331 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "4674795d-7ada-40eb-93d6-524c7af53ee3" (UID: "4674795d-7ada-40eb-93d6-524c7af53ee3"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.314511 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4674795d-7ada-40eb-93d6-524c7af53ee3" (UID: "4674795d-7ada-40eb-93d6-524c7af53ee3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.315447 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "4674795d-7ada-40eb-93d6-524c7af53ee3" (UID: "4674795d-7ada-40eb-93d6-524c7af53ee3"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.333818 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-inventory" (OuterVolumeSpecName: "inventory") pod "4674795d-7ada-40eb-93d6-524c7af53ee3" (UID: "4674795d-7ada-40eb-93d6-524c7af53ee3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.341553 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "4674795d-7ada-40eb-93d6-524c7af53ee3" (UID: "4674795d-7ada-40eb-93d6-524c7af53ee3"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.387194 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfzh2\" (UniqueName: \"kubernetes.io/projected/4674795d-7ada-40eb-93d6-524c7af53ee3-kube-api-access-sfzh2\") on node \"crc\" DevicePath \"\"" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.387228 4745 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.387239 4745 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.387248 4745 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.387259 4745 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.387270 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4674795d-7ada-40eb-93d6-524c7af53ee3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.678032 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" event={"ID":"4674795d-7ada-40eb-93d6-524c7af53ee3","Type":"ContainerDied","Data":"0bcf1fd13a4a67b92dac799531a1b5d24ffa3241adf05f255983bf3eec79a037"} Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.678673 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bcf1fd13a4a67b92dac799531a1b5d24ffa3241adf05f255983bf3eec79a037" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.678108 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.827308 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg"] Oct 10 13:51:13 crc kubenswrapper[4745]: E1010 13:51:13.828075 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4674795d-7ada-40eb-93d6-524c7af53ee3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.828108 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="4674795d-7ada-40eb-93d6-524c7af53ee3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.828461 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="4674795d-7ada-40eb-93d6-524c7af53ee3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.829624 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.832067 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.832493 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.832678 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.834280 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.835196 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:51:13 crc kubenswrapper[4745]: I1010 13:51:13.845976 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg"] Oct 10 13:51:14 crc kubenswrapper[4745]: I1010 13:51:14.001589 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:51:14 crc kubenswrapper[4745]: I1010 13:51:14.001925 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:51:14 crc kubenswrapper[4745]: I1010 13:51:14.001966 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:51:14 crc kubenswrapper[4745]: I1010 13:51:14.002050 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w7t2\" (UniqueName: \"kubernetes.io/projected/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-kube-api-access-8w7t2\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:51:14 crc kubenswrapper[4745]: I1010 13:51:14.002093 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:51:14 crc kubenswrapper[4745]: I1010 13:51:14.103644 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w7t2\" (UniqueName: \"kubernetes.io/projected/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-kube-api-access-8w7t2\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:51:14 crc kubenswrapper[4745]: I1010 13:51:14.103707 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:51:14 crc kubenswrapper[4745]: I1010 13:51:14.103796 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:51:14 crc kubenswrapper[4745]: I1010 13:51:14.103837 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:51:14 crc kubenswrapper[4745]: I1010 13:51:14.103872 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:51:14 crc kubenswrapper[4745]: I1010 13:51:14.110252 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:51:14 crc kubenswrapper[4745]: I1010 13:51:14.111861 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:51:14 crc kubenswrapper[4745]: I1010 13:51:14.112775 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:51:14 crc kubenswrapper[4745]: I1010 13:51:14.113091 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:51:14 crc kubenswrapper[4745]: I1010 13:51:14.124724 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w7t2\" (UniqueName: \"kubernetes.io/projected/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-kube-api-access-8w7t2\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:51:14 crc kubenswrapper[4745]: I1010 13:51:14.149069 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:51:14 crc kubenswrapper[4745]: I1010 13:51:14.723000 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg"] Oct 10 13:51:15 crc kubenswrapper[4745]: I1010 13:51:15.706257 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" event={"ID":"4709df73-fa0c-4b05-aa37-8d32a4aa52f1","Type":"ContainerStarted","Data":"08ac879e0a697b88872fc8baf5b65bc74f3bb1959e40a96f1b84cb22be41d537"} Oct 10 13:51:15 crc kubenswrapper[4745]: I1010 13:51:15.707507 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" event={"ID":"4709df73-fa0c-4b05-aa37-8d32a4aa52f1","Type":"ContainerStarted","Data":"cd596187c6227fbb131960627f3c8c4135c9ad107def65fd46333371771e0692"} Oct 10 13:51:15 crc kubenswrapper[4745]: I1010 13:51:15.740236 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" podStartSLOduration=2.264353546 podStartE2EDuration="2.740206457s" podCreationTimestamp="2025-10-10 13:51:13 +0000 UTC" firstStartedPulling="2025-10-10 13:51:14.743372805 +0000 UTC m=+1988.641029568" lastFinishedPulling="2025-10-10 13:51:15.219225716 +0000 UTC m=+1989.116882479" observedRunningTime="2025-10-10 13:51:15.728119911 +0000 UTC m=+1989.625776714" watchObservedRunningTime="2025-10-10 13:51:15.740206457 +0000 UTC m=+1989.637863260" Oct 10 13:51:16 crc kubenswrapper[4745]: I1010 13:51:16.187658 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:51:16 crc kubenswrapper[4745]: I1010 13:51:16.187834 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:51:16 crc kubenswrapper[4745]: I1010 13:51:16.187928 4745 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:51:16 crc kubenswrapper[4745]: I1010 13:51:16.189787 4745 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ca4a6f66c74821d4fbc4213650a195182b8af7f523a58adff73b04160e45bbe7"} pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 13:51:16 crc kubenswrapper[4745]: I1010 13:51:16.189945 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" containerID="cri-o://ca4a6f66c74821d4fbc4213650a195182b8af7f523a58adff73b04160e45bbe7" gracePeriod=600 Oct 10 13:51:16 crc kubenswrapper[4745]: I1010 13:51:16.720835 4745 generic.go:334] "Generic (PLEG): container finished" podID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerID="ca4a6f66c74821d4fbc4213650a195182b8af7f523a58adff73b04160e45bbe7" exitCode=0 Oct 10 13:51:16 crc kubenswrapper[4745]: I1010 13:51:16.720921 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerDied","Data":"ca4a6f66c74821d4fbc4213650a195182b8af7f523a58adff73b04160e45bbe7"} Oct 10 13:51:16 crc kubenswrapper[4745]: I1010 13:51:16.722231 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9"} Oct 10 13:51:16 crc kubenswrapper[4745]: I1010 13:51:16.722294 4745 scope.go:117] "RemoveContainer" containerID="59312dc2569071d311db229036235d4d7c1d1530d5ad527cdde00b762f5f6998" Oct 10 13:51:27 crc kubenswrapper[4745]: I1010 13:51:27.158109 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ds9bt"] Oct 10 13:51:27 crc kubenswrapper[4745]: I1010 13:51:27.164201 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ds9bt" Oct 10 13:51:27 crc kubenswrapper[4745]: I1010 13:51:27.189878 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ds9bt"] Oct 10 13:51:27 crc kubenswrapper[4745]: I1010 13:51:27.298536 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqfln\" (UniqueName: \"kubernetes.io/projected/aa419d9d-4ba8-4238-b64b-11bc3234f49e-kube-api-access-lqfln\") pod \"redhat-operators-ds9bt\" (UID: \"aa419d9d-4ba8-4238-b64b-11bc3234f49e\") " pod="openshift-marketplace/redhat-operators-ds9bt" Oct 10 13:51:27 crc kubenswrapper[4745]: I1010 13:51:27.298580 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa419d9d-4ba8-4238-b64b-11bc3234f49e-catalog-content\") pod \"redhat-operators-ds9bt\" (UID: \"aa419d9d-4ba8-4238-b64b-11bc3234f49e\") " pod="openshift-marketplace/redhat-operators-ds9bt" Oct 10 13:51:27 crc kubenswrapper[4745]: I1010 13:51:27.298611 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa419d9d-4ba8-4238-b64b-11bc3234f49e-utilities\") pod \"redhat-operators-ds9bt\" (UID: \"aa419d9d-4ba8-4238-b64b-11bc3234f49e\") " pod="openshift-marketplace/redhat-operators-ds9bt" Oct 10 13:51:27 crc kubenswrapper[4745]: I1010 13:51:27.400024 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqfln\" (UniqueName: \"kubernetes.io/projected/aa419d9d-4ba8-4238-b64b-11bc3234f49e-kube-api-access-lqfln\") pod \"redhat-operators-ds9bt\" (UID: \"aa419d9d-4ba8-4238-b64b-11bc3234f49e\") " pod="openshift-marketplace/redhat-operators-ds9bt" Oct 10 13:51:27 crc kubenswrapper[4745]: I1010 13:51:27.400071 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa419d9d-4ba8-4238-b64b-11bc3234f49e-catalog-content\") pod \"redhat-operators-ds9bt\" (UID: \"aa419d9d-4ba8-4238-b64b-11bc3234f49e\") " pod="openshift-marketplace/redhat-operators-ds9bt" Oct 10 13:51:27 crc kubenswrapper[4745]: I1010 13:51:27.400105 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa419d9d-4ba8-4238-b64b-11bc3234f49e-utilities\") pod \"redhat-operators-ds9bt\" (UID: \"aa419d9d-4ba8-4238-b64b-11bc3234f49e\") " pod="openshift-marketplace/redhat-operators-ds9bt" Oct 10 13:51:27 crc kubenswrapper[4745]: I1010 13:51:27.400663 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa419d9d-4ba8-4238-b64b-11bc3234f49e-utilities\") pod \"redhat-operators-ds9bt\" (UID: \"aa419d9d-4ba8-4238-b64b-11bc3234f49e\") " pod="openshift-marketplace/redhat-operators-ds9bt" Oct 10 13:51:27 crc kubenswrapper[4745]: I1010 13:51:27.400767 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa419d9d-4ba8-4238-b64b-11bc3234f49e-catalog-content\") pod \"redhat-operators-ds9bt\" (UID: \"aa419d9d-4ba8-4238-b64b-11bc3234f49e\") " pod="openshift-marketplace/redhat-operators-ds9bt" Oct 10 13:51:27 crc kubenswrapper[4745]: I1010 13:51:27.426862 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqfln\" (UniqueName: \"kubernetes.io/projected/aa419d9d-4ba8-4238-b64b-11bc3234f49e-kube-api-access-lqfln\") pod \"redhat-operators-ds9bt\" (UID: \"aa419d9d-4ba8-4238-b64b-11bc3234f49e\") " pod="openshift-marketplace/redhat-operators-ds9bt" Oct 10 13:51:27 crc kubenswrapper[4745]: I1010 13:51:27.533643 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ds9bt" Oct 10 13:51:27 crc kubenswrapper[4745]: I1010 13:51:27.989014 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ds9bt"] Oct 10 13:51:27 crc kubenswrapper[4745]: W1010 13:51:27.993069 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa419d9d_4ba8_4238_b64b_11bc3234f49e.slice/crio-53f2cdba6947610f719cce8bdf73e5ea0ed6c6a31660d89f3f02d0dfddf59517 WatchSource:0}: Error finding container 53f2cdba6947610f719cce8bdf73e5ea0ed6c6a31660d89f3f02d0dfddf59517: Status 404 returned error can't find the container with id 53f2cdba6947610f719cce8bdf73e5ea0ed6c6a31660d89f3f02d0dfddf59517 Oct 10 13:51:28 crc kubenswrapper[4745]: I1010 13:51:28.843824 4745 generic.go:334] "Generic (PLEG): container finished" podID="aa419d9d-4ba8-4238-b64b-11bc3234f49e" containerID="9ff0fcbf7bee98fb8b2330a0e46046e2673650eabc45c60ecb2552c7815c7925" exitCode=0 Oct 10 13:51:28 crc kubenswrapper[4745]: I1010 13:51:28.843895 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ds9bt" event={"ID":"aa419d9d-4ba8-4238-b64b-11bc3234f49e","Type":"ContainerDied","Data":"9ff0fcbf7bee98fb8b2330a0e46046e2673650eabc45c60ecb2552c7815c7925"} Oct 10 13:51:28 crc kubenswrapper[4745]: I1010 13:51:28.845536 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ds9bt" event={"ID":"aa419d9d-4ba8-4238-b64b-11bc3234f49e","Type":"ContainerStarted","Data":"53f2cdba6947610f719cce8bdf73e5ea0ed6c6a31660d89f3f02d0dfddf59517"} Oct 10 13:51:29 crc kubenswrapper[4745]: I1010 13:51:29.858322 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ds9bt" event={"ID":"aa419d9d-4ba8-4238-b64b-11bc3234f49e","Type":"ContainerStarted","Data":"051a617448468c6c380d6ed860d15595f30a48c9f77b14f9a4730530fac9238a"} Oct 10 13:51:30 crc kubenswrapper[4745]: I1010 13:51:30.878209 4745 generic.go:334] "Generic (PLEG): container finished" podID="aa419d9d-4ba8-4238-b64b-11bc3234f49e" containerID="051a617448468c6c380d6ed860d15595f30a48c9f77b14f9a4730530fac9238a" exitCode=0 Oct 10 13:51:30 crc kubenswrapper[4745]: I1010 13:51:30.878973 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ds9bt" event={"ID":"aa419d9d-4ba8-4238-b64b-11bc3234f49e","Type":"ContainerDied","Data":"051a617448468c6c380d6ed860d15595f30a48c9f77b14f9a4730530fac9238a"} Oct 10 13:51:32 crc kubenswrapper[4745]: I1010 13:51:32.909207 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ds9bt" event={"ID":"aa419d9d-4ba8-4238-b64b-11bc3234f49e","Type":"ContainerStarted","Data":"36c5b72321593478397150bf9e57abe41e39a29e839c6191cb2906d99f7629ec"} Oct 10 13:51:32 crc kubenswrapper[4745]: I1010 13:51:32.938424 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ds9bt" podStartSLOduration=3.468185551 podStartE2EDuration="5.938400273s" podCreationTimestamp="2025-10-10 13:51:27 +0000 UTC" firstStartedPulling="2025-10-10 13:51:28.845246164 +0000 UTC m=+2002.742902927" lastFinishedPulling="2025-10-10 13:51:31.315460846 +0000 UTC m=+2005.213117649" observedRunningTime="2025-10-10 13:51:32.931013317 +0000 UTC m=+2006.828670090" watchObservedRunningTime="2025-10-10 13:51:32.938400273 +0000 UTC m=+2006.836057036" Oct 10 13:51:37 crc kubenswrapper[4745]: I1010 13:51:37.534414 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ds9bt" Oct 10 13:51:37 crc kubenswrapper[4745]: I1010 13:51:37.536879 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ds9bt" Oct 10 13:51:37 crc kubenswrapper[4745]: I1010 13:51:37.646157 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ds9bt" Oct 10 13:51:38 crc kubenswrapper[4745]: I1010 13:51:38.000236 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ds9bt" Oct 10 13:51:38 crc kubenswrapper[4745]: I1010 13:51:38.042018 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ds9bt"] Oct 10 13:51:39 crc kubenswrapper[4745]: I1010 13:51:39.980525 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ds9bt" podUID="aa419d9d-4ba8-4238-b64b-11bc3234f49e" containerName="registry-server" containerID="cri-o://36c5b72321593478397150bf9e57abe41e39a29e839c6191cb2906d99f7629ec" gracePeriod=2 Oct 10 13:51:40 crc kubenswrapper[4745]: I1010 13:51:40.434178 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ds9bt" Oct 10 13:51:40 crc kubenswrapper[4745]: I1010 13:51:40.562500 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqfln\" (UniqueName: \"kubernetes.io/projected/aa419d9d-4ba8-4238-b64b-11bc3234f49e-kube-api-access-lqfln\") pod \"aa419d9d-4ba8-4238-b64b-11bc3234f49e\" (UID: \"aa419d9d-4ba8-4238-b64b-11bc3234f49e\") " Oct 10 13:51:40 crc kubenswrapper[4745]: I1010 13:51:40.562600 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa419d9d-4ba8-4238-b64b-11bc3234f49e-catalog-content\") pod \"aa419d9d-4ba8-4238-b64b-11bc3234f49e\" (UID: \"aa419d9d-4ba8-4238-b64b-11bc3234f49e\") " Oct 10 13:51:40 crc kubenswrapper[4745]: I1010 13:51:40.562714 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa419d9d-4ba8-4238-b64b-11bc3234f49e-utilities\") pod \"aa419d9d-4ba8-4238-b64b-11bc3234f49e\" (UID: \"aa419d9d-4ba8-4238-b64b-11bc3234f49e\") " Oct 10 13:51:40 crc kubenswrapper[4745]: I1010 13:51:40.563503 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa419d9d-4ba8-4238-b64b-11bc3234f49e-utilities" (OuterVolumeSpecName: "utilities") pod "aa419d9d-4ba8-4238-b64b-11bc3234f49e" (UID: "aa419d9d-4ba8-4238-b64b-11bc3234f49e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:51:40 crc kubenswrapper[4745]: I1010 13:51:40.567501 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa419d9d-4ba8-4238-b64b-11bc3234f49e-kube-api-access-lqfln" (OuterVolumeSpecName: "kube-api-access-lqfln") pod "aa419d9d-4ba8-4238-b64b-11bc3234f49e" (UID: "aa419d9d-4ba8-4238-b64b-11bc3234f49e"). InnerVolumeSpecName "kube-api-access-lqfln". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:51:40 crc kubenswrapper[4745]: I1010 13:51:40.641510 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa419d9d-4ba8-4238-b64b-11bc3234f49e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa419d9d-4ba8-4238-b64b-11bc3234f49e" (UID: "aa419d9d-4ba8-4238-b64b-11bc3234f49e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:51:40 crc kubenswrapper[4745]: I1010 13:51:40.665126 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqfln\" (UniqueName: \"kubernetes.io/projected/aa419d9d-4ba8-4238-b64b-11bc3234f49e-kube-api-access-lqfln\") on node \"crc\" DevicePath \"\"" Oct 10 13:51:40 crc kubenswrapper[4745]: I1010 13:51:40.665162 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa419d9d-4ba8-4238-b64b-11bc3234f49e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:51:40 crc kubenswrapper[4745]: I1010 13:51:40.665172 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa419d9d-4ba8-4238-b64b-11bc3234f49e-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:51:40 crc kubenswrapper[4745]: I1010 13:51:40.992286 4745 generic.go:334] "Generic (PLEG): container finished" podID="aa419d9d-4ba8-4238-b64b-11bc3234f49e" containerID="36c5b72321593478397150bf9e57abe41e39a29e839c6191cb2906d99f7629ec" exitCode=0 Oct 10 13:51:40 crc kubenswrapper[4745]: I1010 13:51:40.992330 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ds9bt" event={"ID":"aa419d9d-4ba8-4238-b64b-11bc3234f49e","Type":"ContainerDied","Data":"36c5b72321593478397150bf9e57abe41e39a29e839c6191cb2906d99f7629ec"} Oct 10 13:51:40 crc kubenswrapper[4745]: I1010 13:51:40.992353 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ds9bt" Oct 10 13:51:40 crc kubenswrapper[4745]: I1010 13:51:40.992369 4745 scope.go:117] "RemoveContainer" containerID="36c5b72321593478397150bf9e57abe41e39a29e839c6191cb2906d99f7629ec" Oct 10 13:51:40 crc kubenswrapper[4745]: I1010 13:51:40.992358 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ds9bt" event={"ID":"aa419d9d-4ba8-4238-b64b-11bc3234f49e","Type":"ContainerDied","Data":"53f2cdba6947610f719cce8bdf73e5ea0ed6c6a31660d89f3f02d0dfddf59517"} Oct 10 13:51:41 crc kubenswrapper[4745]: I1010 13:51:41.017194 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ds9bt"] Oct 10 13:51:41 crc kubenswrapper[4745]: I1010 13:51:41.020725 4745 scope.go:117] "RemoveContainer" containerID="051a617448468c6c380d6ed860d15595f30a48c9f77b14f9a4730530fac9238a" Oct 10 13:51:41 crc kubenswrapper[4745]: I1010 13:51:41.025408 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ds9bt"] Oct 10 13:51:41 crc kubenswrapper[4745]: I1010 13:51:41.042149 4745 scope.go:117] "RemoveContainer" containerID="9ff0fcbf7bee98fb8b2330a0e46046e2673650eabc45c60ecb2552c7815c7925" Oct 10 13:51:41 crc kubenswrapper[4745]: I1010 13:51:41.097154 4745 scope.go:117] "RemoveContainer" containerID="36c5b72321593478397150bf9e57abe41e39a29e839c6191cb2906d99f7629ec" Oct 10 13:51:41 crc kubenswrapper[4745]: E1010 13:51:41.097743 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36c5b72321593478397150bf9e57abe41e39a29e839c6191cb2906d99f7629ec\": container with ID starting with 36c5b72321593478397150bf9e57abe41e39a29e839c6191cb2906d99f7629ec not found: ID does not exist" containerID="36c5b72321593478397150bf9e57abe41e39a29e839c6191cb2906d99f7629ec" Oct 10 13:51:41 crc kubenswrapper[4745]: I1010 13:51:41.097789 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36c5b72321593478397150bf9e57abe41e39a29e839c6191cb2906d99f7629ec"} err="failed to get container status \"36c5b72321593478397150bf9e57abe41e39a29e839c6191cb2906d99f7629ec\": rpc error: code = NotFound desc = could not find container \"36c5b72321593478397150bf9e57abe41e39a29e839c6191cb2906d99f7629ec\": container with ID starting with 36c5b72321593478397150bf9e57abe41e39a29e839c6191cb2906d99f7629ec not found: ID does not exist" Oct 10 13:51:41 crc kubenswrapper[4745]: I1010 13:51:41.097815 4745 scope.go:117] "RemoveContainer" containerID="051a617448468c6c380d6ed860d15595f30a48c9f77b14f9a4730530fac9238a" Oct 10 13:51:41 crc kubenswrapper[4745]: E1010 13:51:41.098219 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"051a617448468c6c380d6ed860d15595f30a48c9f77b14f9a4730530fac9238a\": container with ID starting with 051a617448468c6c380d6ed860d15595f30a48c9f77b14f9a4730530fac9238a not found: ID does not exist" containerID="051a617448468c6c380d6ed860d15595f30a48c9f77b14f9a4730530fac9238a" Oct 10 13:51:41 crc kubenswrapper[4745]: I1010 13:51:41.098253 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"051a617448468c6c380d6ed860d15595f30a48c9f77b14f9a4730530fac9238a"} err="failed to get container status \"051a617448468c6c380d6ed860d15595f30a48c9f77b14f9a4730530fac9238a\": rpc error: code = NotFound desc = could not find container \"051a617448468c6c380d6ed860d15595f30a48c9f77b14f9a4730530fac9238a\": container with ID starting with 051a617448468c6c380d6ed860d15595f30a48c9f77b14f9a4730530fac9238a not found: ID does not exist" Oct 10 13:51:41 crc kubenswrapper[4745]: I1010 13:51:41.098275 4745 scope.go:117] "RemoveContainer" containerID="9ff0fcbf7bee98fb8b2330a0e46046e2673650eabc45c60ecb2552c7815c7925" Oct 10 13:51:41 crc kubenswrapper[4745]: E1010 13:51:41.098568 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ff0fcbf7bee98fb8b2330a0e46046e2673650eabc45c60ecb2552c7815c7925\": container with ID starting with 9ff0fcbf7bee98fb8b2330a0e46046e2673650eabc45c60ecb2552c7815c7925 not found: ID does not exist" containerID="9ff0fcbf7bee98fb8b2330a0e46046e2673650eabc45c60ecb2552c7815c7925" Oct 10 13:51:41 crc kubenswrapper[4745]: I1010 13:51:41.098626 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ff0fcbf7bee98fb8b2330a0e46046e2673650eabc45c60ecb2552c7815c7925"} err="failed to get container status \"9ff0fcbf7bee98fb8b2330a0e46046e2673650eabc45c60ecb2552c7815c7925\": rpc error: code = NotFound desc = could not find container \"9ff0fcbf7bee98fb8b2330a0e46046e2673650eabc45c60ecb2552c7815c7925\": container with ID starting with 9ff0fcbf7bee98fb8b2330a0e46046e2673650eabc45c60ecb2552c7815c7925 not found: ID does not exist" Oct 10 13:51:42 crc kubenswrapper[4745]: I1010 13:51:42.757117 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa419d9d-4ba8-4238-b64b-11bc3234f49e" path="/var/lib/kubelet/pods/aa419d9d-4ba8-4238-b64b-11bc3234f49e/volumes" Oct 10 13:51:58 crc kubenswrapper[4745]: I1010 13:51:58.344094 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-trmp2"] Oct 10 13:51:58 crc kubenswrapper[4745]: E1010 13:51:58.344971 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa419d9d-4ba8-4238-b64b-11bc3234f49e" containerName="registry-server" Oct 10 13:51:58 crc kubenswrapper[4745]: I1010 13:51:58.344985 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa419d9d-4ba8-4238-b64b-11bc3234f49e" containerName="registry-server" Oct 10 13:51:58 crc kubenswrapper[4745]: E1010 13:51:58.344999 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa419d9d-4ba8-4238-b64b-11bc3234f49e" containerName="extract-utilities" Oct 10 13:51:58 crc kubenswrapper[4745]: I1010 13:51:58.345005 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa419d9d-4ba8-4238-b64b-11bc3234f49e" containerName="extract-utilities" Oct 10 13:51:58 crc kubenswrapper[4745]: E1010 13:51:58.345032 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa419d9d-4ba8-4238-b64b-11bc3234f49e" containerName="extract-content" Oct 10 13:51:58 crc kubenswrapper[4745]: I1010 13:51:58.345038 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa419d9d-4ba8-4238-b64b-11bc3234f49e" containerName="extract-content" Oct 10 13:51:58 crc kubenswrapper[4745]: I1010 13:51:58.345250 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa419d9d-4ba8-4238-b64b-11bc3234f49e" containerName="registry-server" Oct 10 13:51:58 crc kubenswrapper[4745]: I1010 13:51:58.348652 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trmp2" Oct 10 13:51:58 crc kubenswrapper[4745]: I1010 13:51:58.363946 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-trmp2"] Oct 10 13:51:58 crc kubenswrapper[4745]: I1010 13:51:58.436875 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a95002a-e1dd-42a5-9375-2743704192c1-utilities\") pod \"redhat-marketplace-trmp2\" (UID: \"9a95002a-e1dd-42a5-9375-2743704192c1\") " pod="openshift-marketplace/redhat-marketplace-trmp2" Oct 10 13:51:58 crc kubenswrapper[4745]: I1010 13:51:58.436940 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a95002a-e1dd-42a5-9375-2743704192c1-catalog-content\") pod \"redhat-marketplace-trmp2\" (UID: \"9a95002a-e1dd-42a5-9375-2743704192c1\") " pod="openshift-marketplace/redhat-marketplace-trmp2" Oct 10 13:51:58 crc kubenswrapper[4745]: I1010 13:51:58.437086 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s8r5\" (UniqueName: \"kubernetes.io/projected/9a95002a-e1dd-42a5-9375-2743704192c1-kube-api-access-6s8r5\") pod \"redhat-marketplace-trmp2\" (UID: \"9a95002a-e1dd-42a5-9375-2743704192c1\") " pod="openshift-marketplace/redhat-marketplace-trmp2" Oct 10 13:51:58 crc kubenswrapper[4745]: I1010 13:51:58.539120 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a95002a-e1dd-42a5-9375-2743704192c1-utilities\") pod \"redhat-marketplace-trmp2\" (UID: \"9a95002a-e1dd-42a5-9375-2743704192c1\") " pod="openshift-marketplace/redhat-marketplace-trmp2" Oct 10 13:51:58 crc kubenswrapper[4745]: I1010 13:51:58.539189 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a95002a-e1dd-42a5-9375-2743704192c1-catalog-content\") pod \"redhat-marketplace-trmp2\" (UID: \"9a95002a-e1dd-42a5-9375-2743704192c1\") " pod="openshift-marketplace/redhat-marketplace-trmp2" Oct 10 13:51:58 crc kubenswrapper[4745]: I1010 13:51:58.539229 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s8r5\" (UniqueName: \"kubernetes.io/projected/9a95002a-e1dd-42a5-9375-2743704192c1-kube-api-access-6s8r5\") pod \"redhat-marketplace-trmp2\" (UID: \"9a95002a-e1dd-42a5-9375-2743704192c1\") " pod="openshift-marketplace/redhat-marketplace-trmp2" Oct 10 13:51:58 crc kubenswrapper[4745]: I1010 13:51:58.539674 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a95002a-e1dd-42a5-9375-2743704192c1-utilities\") pod \"redhat-marketplace-trmp2\" (UID: \"9a95002a-e1dd-42a5-9375-2743704192c1\") " pod="openshift-marketplace/redhat-marketplace-trmp2" Oct 10 13:51:58 crc kubenswrapper[4745]: I1010 13:51:58.539721 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a95002a-e1dd-42a5-9375-2743704192c1-catalog-content\") pod \"redhat-marketplace-trmp2\" (UID: \"9a95002a-e1dd-42a5-9375-2743704192c1\") " pod="openshift-marketplace/redhat-marketplace-trmp2" Oct 10 13:51:58 crc kubenswrapper[4745]: I1010 13:51:58.561070 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s8r5\" (UniqueName: \"kubernetes.io/projected/9a95002a-e1dd-42a5-9375-2743704192c1-kube-api-access-6s8r5\") pod \"redhat-marketplace-trmp2\" (UID: \"9a95002a-e1dd-42a5-9375-2743704192c1\") " pod="openshift-marketplace/redhat-marketplace-trmp2" Oct 10 13:51:58 crc kubenswrapper[4745]: I1010 13:51:58.677781 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trmp2" Oct 10 13:51:59 crc kubenswrapper[4745]: I1010 13:51:59.166711 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-trmp2"] Oct 10 13:52:00 crc kubenswrapper[4745]: I1010 13:52:00.189036 4745 generic.go:334] "Generic (PLEG): container finished" podID="9a95002a-e1dd-42a5-9375-2743704192c1" containerID="98e26564b6ca97814b98174ad0dae19660075a915738707bd1307dde0f8533d1" exitCode=0 Oct 10 13:52:00 crc kubenswrapper[4745]: I1010 13:52:00.189111 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trmp2" event={"ID":"9a95002a-e1dd-42a5-9375-2743704192c1","Type":"ContainerDied","Data":"98e26564b6ca97814b98174ad0dae19660075a915738707bd1307dde0f8533d1"} Oct 10 13:52:00 crc kubenswrapper[4745]: I1010 13:52:00.189493 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trmp2" event={"ID":"9a95002a-e1dd-42a5-9375-2743704192c1","Type":"ContainerStarted","Data":"b2f1f2ff7ea337b58e4ee0b54e95c51d1f1cd56599fa1b8cfeeb0a5ca1990fc4"} Oct 10 13:52:02 crc kubenswrapper[4745]: I1010 13:52:02.213866 4745 generic.go:334] "Generic (PLEG): container finished" podID="9a95002a-e1dd-42a5-9375-2743704192c1" containerID="243171639ccf7984c056aee7be94c2fc789025f61f238ba6f07e272571ce9b2a" exitCode=0 Oct 10 13:52:02 crc kubenswrapper[4745]: I1010 13:52:02.213941 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trmp2" event={"ID":"9a95002a-e1dd-42a5-9375-2743704192c1","Type":"ContainerDied","Data":"243171639ccf7984c056aee7be94c2fc789025f61f238ba6f07e272571ce9b2a"} Oct 10 13:52:03 crc kubenswrapper[4745]: I1010 13:52:03.225593 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trmp2" event={"ID":"9a95002a-e1dd-42a5-9375-2743704192c1","Type":"ContainerStarted","Data":"32c6584bc11f27cbff1b6f472d370e2e02a6d86bb78a920117a798dd3220dd4a"} Oct 10 13:52:03 crc kubenswrapper[4745]: I1010 13:52:03.248092 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-trmp2" podStartSLOduration=2.791511902 podStartE2EDuration="5.248068109s" podCreationTimestamp="2025-10-10 13:51:58 +0000 UTC" firstStartedPulling="2025-10-10 13:52:00.191919085 +0000 UTC m=+2034.089575858" lastFinishedPulling="2025-10-10 13:52:02.648475292 +0000 UTC m=+2036.546132065" observedRunningTime="2025-10-10 13:52:03.243457512 +0000 UTC m=+2037.141114285" watchObservedRunningTime="2025-10-10 13:52:03.248068109 +0000 UTC m=+2037.145724882" Oct 10 13:52:08 crc kubenswrapper[4745]: I1010 13:52:08.678287 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-trmp2" Oct 10 13:52:08 crc kubenswrapper[4745]: I1010 13:52:08.678892 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-trmp2" Oct 10 13:52:08 crc kubenswrapper[4745]: I1010 13:52:08.733434 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-trmp2" Oct 10 13:52:09 crc kubenswrapper[4745]: I1010 13:52:09.362672 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-trmp2" Oct 10 13:52:09 crc kubenswrapper[4745]: I1010 13:52:09.440184 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-trmp2"] Oct 10 13:52:10 crc kubenswrapper[4745]: I1010 13:52:10.786834 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gk5tk"] Oct 10 13:52:10 crc kubenswrapper[4745]: I1010 13:52:10.791586 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gk5tk" Oct 10 13:52:10 crc kubenswrapper[4745]: I1010 13:52:10.804491 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gk5tk"] Oct 10 13:52:10 crc kubenswrapper[4745]: I1010 13:52:10.913670 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8514ec89-60e8-471b-9476-80f9255695d4-catalog-content\") pod \"certified-operators-gk5tk\" (UID: \"8514ec89-60e8-471b-9476-80f9255695d4\") " pod="openshift-marketplace/certified-operators-gk5tk" Oct 10 13:52:10 crc kubenswrapper[4745]: I1010 13:52:10.914068 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8514ec89-60e8-471b-9476-80f9255695d4-utilities\") pod \"certified-operators-gk5tk\" (UID: \"8514ec89-60e8-471b-9476-80f9255695d4\") " pod="openshift-marketplace/certified-operators-gk5tk" Oct 10 13:52:10 crc kubenswrapper[4745]: I1010 13:52:10.914235 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6dg2\" (UniqueName: \"kubernetes.io/projected/8514ec89-60e8-471b-9476-80f9255695d4-kube-api-access-t6dg2\") pod \"certified-operators-gk5tk\" (UID: \"8514ec89-60e8-471b-9476-80f9255695d4\") " pod="openshift-marketplace/certified-operators-gk5tk" Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.015691 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8514ec89-60e8-471b-9476-80f9255695d4-utilities\") pod \"certified-operators-gk5tk\" (UID: \"8514ec89-60e8-471b-9476-80f9255695d4\") " pod="openshift-marketplace/certified-operators-gk5tk" Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.015783 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6dg2\" (UniqueName: \"kubernetes.io/projected/8514ec89-60e8-471b-9476-80f9255695d4-kube-api-access-t6dg2\") pod \"certified-operators-gk5tk\" (UID: \"8514ec89-60e8-471b-9476-80f9255695d4\") " pod="openshift-marketplace/certified-operators-gk5tk" Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.015902 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8514ec89-60e8-471b-9476-80f9255695d4-catalog-content\") pod \"certified-operators-gk5tk\" (UID: \"8514ec89-60e8-471b-9476-80f9255695d4\") " pod="openshift-marketplace/certified-operators-gk5tk" Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.016419 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8514ec89-60e8-471b-9476-80f9255695d4-utilities\") pod \"certified-operators-gk5tk\" (UID: \"8514ec89-60e8-471b-9476-80f9255695d4\") " pod="openshift-marketplace/certified-operators-gk5tk" Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.016455 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8514ec89-60e8-471b-9476-80f9255695d4-catalog-content\") pod \"certified-operators-gk5tk\" (UID: \"8514ec89-60e8-471b-9476-80f9255695d4\") " pod="openshift-marketplace/certified-operators-gk5tk" Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.039351 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6dg2\" (UniqueName: \"kubernetes.io/projected/8514ec89-60e8-471b-9476-80f9255695d4-kube-api-access-t6dg2\") pod \"certified-operators-gk5tk\" (UID: \"8514ec89-60e8-471b-9476-80f9255695d4\") " pod="openshift-marketplace/certified-operators-gk5tk" Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.113131 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gk5tk" Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.327853 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-trmp2" podUID="9a95002a-e1dd-42a5-9375-2743704192c1" containerName="registry-server" containerID="cri-o://32c6584bc11f27cbff1b6f472d370e2e02a6d86bb78a920117a798dd3220dd4a" gracePeriod=2 Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.653415 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gk5tk"] Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.771271 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trmp2" Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.830781 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a95002a-e1dd-42a5-9375-2743704192c1-catalog-content\") pod \"9a95002a-e1dd-42a5-9375-2743704192c1\" (UID: \"9a95002a-e1dd-42a5-9375-2743704192c1\") " Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.830915 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a95002a-e1dd-42a5-9375-2743704192c1-utilities\") pod \"9a95002a-e1dd-42a5-9375-2743704192c1\" (UID: \"9a95002a-e1dd-42a5-9375-2743704192c1\") " Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.831106 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6s8r5\" (UniqueName: \"kubernetes.io/projected/9a95002a-e1dd-42a5-9375-2743704192c1-kube-api-access-6s8r5\") pod \"9a95002a-e1dd-42a5-9375-2743704192c1\" (UID: \"9a95002a-e1dd-42a5-9375-2743704192c1\") " Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.834101 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a95002a-e1dd-42a5-9375-2743704192c1-utilities" (OuterVolumeSpecName: "utilities") pod "9a95002a-e1dd-42a5-9375-2743704192c1" (UID: "9a95002a-e1dd-42a5-9375-2743704192c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.838954 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a95002a-e1dd-42a5-9375-2743704192c1-kube-api-access-6s8r5" (OuterVolumeSpecName: "kube-api-access-6s8r5") pod "9a95002a-e1dd-42a5-9375-2743704192c1" (UID: "9a95002a-e1dd-42a5-9375-2743704192c1"). InnerVolumeSpecName "kube-api-access-6s8r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.843050 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a95002a-e1dd-42a5-9375-2743704192c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a95002a-e1dd-42a5-9375-2743704192c1" (UID: "9a95002a-e1dd-42a5-9375-2743704192c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.932684 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a95002a-e1dd-42a5-9375-2743704192c1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.932713 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a95002a-e1dd-42a5-9375-2743704192c1-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:52:11 crc kubenswrapper[4745]: I1010 13:52:11.932723 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6s8r5\" (UniqueName: \"kubernetes.io/projected/9a95002a-e1dd-42a5-9375-2743704192c1-kube-api-access-6s8r5\") on node \"crc\" DevicePath \"\"" Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.339307 4745 generic.go:334] "Generic (PLEG): container finished" podID="8514ec89-60e8-471b-9476-80f9255695d4" containerID="c88e5cfec1c623e5ad69a5475fcdced12cfc88cec7a26adca79c971e60e94a4c" exitCode=0 Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.339472 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gk5tk" event={"ID":"8514ec89-60e8-471b-9476-80f9255695d4","Type":"ContainerDied","Data":"c88e5cfec1c623e5ad69a5475fcdced12cfc88cec7a26adca79c971e60e94a4c"} Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.339797 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gk5tk" event={"ID":"8514ec89-60e8-471b-9476-80f9255695d4","Type":"ContainerStarted","Data":"96ff874e986c5631bfa3f2781d6e8fbcab56062fdece0b71e48bc61244317017"} Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.343064 4745 generic.go:334] "Generic (PLEG): container finished" podID="9a95002a-e1dd-42a5-9375-2743704192c1" containerID="32c6584bc11f27cbff1b6f472d370e2e02a6d86bb78a920117a798dd3220dd4a" exitCode=0 Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.343113 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trmp2" event={"ID":"9a95002a-e1dd-42a5-9375-2743704192c1","Type":"ContainerDied","Data":"32c6584bc11f27cbff1b6f472d370e2e02a6d86bb78a920117a798dd3220dd4a"} Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.343151 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trmp2" event={"ID":"9a95002a-e1dd-42a5-9375-2743704192c1","Type":"ContainerDied","Data":"b2f1f2ff7ea337b58e4ee0b54e95c51d1f1cd56599fa1b8cfeeb0a5ca1990fc4"} Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.343190 4745 scope.go:117] "RemoveContainer" containerID="32c6584bc11f27cbff1b6f472d370e2e02a6d86bb78a920117a798dd3220dd4a" Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.343409 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trmp2" Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.377777 4745 scope.go:117] "RemoveContainer" containerID="243171639ccf7984c056aee7be94c2fc789025f61f238ba6f07e272571ce9b2a" Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.407672 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-trmp2"] Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.420037 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-trmp2"] Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.429642 4745 scope.go:117] "RemoveContainer" containerID="98e26564b6ca97814b98174ad0dae19660075a915738707bd1307dde0f8533d1" Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.463226 4745 scope.go:117] "RemoveContainer" containerID="32c6584bc11f27cbff1b6f472d370e2e02a6d86bb78a920117a798dd3220dd4a" Oct 10 13:52:12 crc kubenswrapper[4745]: E1010 13:52:12.463661 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32c6584bc11f27cbff1b6f472d370e2e02a6d86bb78a920117a798dd3220dd4a\": container with ID starting with 32c6584bc11f27cbff1b6f472d370e2e02a6d86bb78a920117a798dd3220dd4a not found: ID does not exist" containerID="32c6584bc11f27cbff1b6f472d370e2e02a6d86bb78a920117a798dd3220dd4a" Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.463807 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32c6584bc11f27cbff1b6f472d370e2e02a6d86bb78a920117a798dd3220dd4a"} err="failed to get container status \"32c6584bc11f27cbff1b6f472d370e2e02a6d86bb78a920117a798dd3220dd4a\": rpc error: code = NotFound desc = could not find container \"32c6584bc11f27cbff1b6f472d370e2e02a6d86bb78a920117a798dd3220dd4a\": container with ID starting with 32c6584bc11f27cbff1b6f472d370e2e02a6d86bb78a920117a798dd3220dd4a not found: ID does not exist" Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.463846 4745 scope.go:117] "RemoveContainer" containerID="243171639ccf7984c056aee7be94c2fc789025f61f238ba6f07e272571ce9b2a" Oct 10 13:52:12 crc kubenswrapper[4745]: E1010 13:52:12.464485 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"243171639ccf7984c056aee7be94c2fc789025f61f238ba6f07e272571ce9b2a\": container with ID starting with 243171639ccf7984c056aee7be94c2fc789025f61f238ba6f07e272571ce9b2a not found: ID does not exist" containerID="243171639ccf7984c056aee7be94c2fc789025f61f238ba6f07e272571ce9b2a" Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.464531 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"243171639ccf7984c056aee7be94c2fc789025f61f238ba6f07e272571ce9b2a"} err="failed to get container status \"243171639ccf7984c056aee7be94c2fc789025f61f238ba6f07e272571ce9b2a\": rpc error: code = NotFound desc = could not find container \"243171639ccf7984c056aee7be94c2fc789025f61f238ba6f07e272571ce9b2a\": container with ID starting with 243171639ccf7984c056aee7be94c2fc789025f61f238ba6f07e272571ce9b2a not found: ID does not exist" Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.464563 4745 scope.go:117] "RemoveContainer" containerID="98e26564b6ca97814b98174ad0dae19660075a915738707bd1307dde0f8533d1" Oct 10 13:52:12 crc kubenswrapper[4745]: E1010 13:52:12.465277 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98e26564b6ca97814b98174ad0dae19660075a915738707bd1307dde0f8533d1\": container with ID starting with 98e26564b6ca97814b98174ad0dae19660075a915738707bd1307dde0f8533d1 not found: ID does not exist" containerID="98e26564b6ca97814b98174ad0dae19660075a915738707bd1307dde0f8533d1" Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.465307 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98e26564b6ca97814b98174ad0dae19660075a915738707bd1307dde0f8533d1"} err="failed to get container status \"98e26564b6ca97814b98174ad0dae19660075a915738707bd1307dde0f8533d1\": rpc error: code = NotFound desc = could not find container \"98e26564b6ca97814b98174ad0dae19660075a915738707bd1307dde0f8533d1\": container with ID starting with 98e26564b6ca97814b98174ad0dae19660075a915738707bd1307dde0f8533d1 not found: ID does not exist" Oct 10 13:52:12 crc kubenswrapper[4745]: I1010 13:52:12.760212 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a95002a-e1dd-42a5-9375-2743704192c1" path="/var/lib/kubelet/pods/9a95002a-e1dd-42a5-9375-2743704192c1/volumes" Oct 10 13:52:14 crc kubenswrapper[4745]: I1010 13:52:14.368995 4745 generic.go:334] "Generic (PLEG): container finished" podID="8514ec89-60e8-471b-9476-80f9255695d4" containerID="0672f61add8e9f036084c6bfc5ec58214c61b1dfd98cf47ecde878ddf68191ad" exitCode=0 Oct 10 13:52:14 crc kubenswrapper[4745]: I1010 13:52:14.369241 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gk5tk" event={"ID":"8514ec89-60e8-471b-9476-80f9255695d4","Type":"ContainerDied","Data":"0672f61add8e9f036084c6bfc5ec58214c61b1dfd98cf47ecde878ddf68191ad"} Oct 10 13:52:15 crc kubenswrapper[4745]: I1010 13:52:15.380856 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gk5tk" event={"ID":"8514ec89-60e8-471b-9476-80f9255695d4","Type":"ContainerStarted","Data":"407299999e0652c4be91e60317e1d828fa2df51ebb733566f6f3d178902139f3"} Oct 10 13:52:15 crc kubenswrapper[4745]: I1010 13:52:15.413853 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gk5tk" podStartSLOduration=2.907743088 podStartE2EDuration="5.413834407s" podCreationTimestamp="2025-10-10 13:52:10 +0000 UTC" firstStartedPulling="2025-10-10 13:52:12.345049644 +0000 UTC m=+2046.242706417" lastFinishedPulling="2025-10-10 13:52:14.851140963 +0000 UTC m=+2048.748797736" observedRunningTime="2025-10-10 13:52:15.406228366 +0000 UTC m=+2049.303885149" watchObservedRunningTime="2025-10-10 13:52:15.413834407 +0000 UTC m=+2049.311491190" Oct 10 13:52:21 crc kubenswrapper[4745]: I1010 13:52:21.113803 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gk5tk" Oct 10 13:52:21 crc kubenswrapper[4745]: I1010 13:52:21.114663 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gk5tk" Oct 10 13:52:21 crc kubenswrapper[4745]: I1010 13:52:21.183908 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gk5tk" Oct 10 13:52:21 crc kubenswrapper[4745]: I1010 13:52:21.523034 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gk5tk" Oct 10 13:52:21 crc kubenswrapper[4745]: I1010 13:52:21.600291 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gk5tk"] Oct 10 13:52:23 crc kubenswrapper[4745]: I1010 13:52:23.480229 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gk5tk" podUID="8514ec89-60e8-471b-9476-80f9255695d4" containerName="registry-server" containerID="cri-o://407299999e0652c4be91e60317e1d828fa2df51ebb733566f6f3d178902139f3" gracePeriod=2 Oct 10 13:52:23 crc kubenswrapper[4745]: I1010 13:52:23.996521 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gk5tk" Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.144631 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8514ec89-60e8-471b-9476-80f9255695d4-utilities\") pod \"8514ec89-60e8-471b-9476-80f9255695d4\" (UID: \"8514ec89-60e8-471b-9476-80f9255695d4\") " Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.145126 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8514ec89-60e8-471b-9476-80f9255695d4-catalog-content\") pod \"8514ec89-60e8-471b-9476-80f9255695d4\" (UID: \"8514ec89-60e8-471b-9476-80f9255695d4\") " Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.145435 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6dg2\" (UniqueName: \"kubernetes.io/projected/8514ec89-60e8-471b-9476-80f9255695d4-kube-api-access-t6dg2\") pod \"8514ec89-60e8-471b-9476-80f9255695d4\" (UID: \"8514ec89-60e8-471b-9476-80f9255695d4\") " Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.145801 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8514ec89-60e8-471b-9476-80f9255695d4-utilities" (OuterVolumeSpecName: "utilities") pod "8514ec89-60e8-471b-9476-80f9255695d4" (UID: "8514ec89-60e8-471b-9476-80f9255695d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.146447 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8514ec89-60e8-471b-9476-80f9255695d4-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.154529 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8514ec89-60e8-471b-9476-80f9255695d4-kube-api-access-t6dg2" (OuterVolumeSpecName: "kube-api-access-t6dg2") pod "8514ec89-60e8-471b-9476-80f9255695d4" (UID: "8514ec89-60e8-471b-9476-80f9255695d4"). InnerVolumeSpecName "kube-api-access-t6dg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.249160 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6dg2\" (UniqueName: \"kubernetes.io/projected/8514ec89-60e8-471b-9476-80f9255695d4-kube-api-access-t6dg2\") on node \"crc\" DevicePath \"\"" Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.497111 4745 generic.go:334] "Generic (PLEG): container finished" podID="8514ec89-60e8-471b-9476-80f9255695d4" containerID="407299999e0652c4be91e60317e1d828fa2df51ebb733566f6f3d178902139f3" exitCode=0 Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.497154 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gk5tk" event={"ID":"8514ec89-60e8-471b-9476-80f9255695d4","Type":"ContainerDied","Data":"407299999e0652c4be91e60317e1d828fa2df51ebb733566f6f3d178902139f3"} Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.497184 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gk5tk" event={"ID":"8514ec89-60e8-471b-9476-80f9255695d4","Type":"ContainerDied","Data":"96ff874e986c5631bfa3f2781d6e8fbcab56062fdece0b71e48bc61244317017"} Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.497203 4745 scope.go:117] "RemoveContainer" containerID="407299999e0652c4be91e60317e1d828fa2df51ebb733566f6f3d178902139f3" Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.497335 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gk5tk" Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.528782 4745 scope.go:117] "RemoveContainer" containerID="0672f61add8e9f036084c6bfc5ec58214c61b1dfd98cf47ecde878ddf68191ad" Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.561652 4745 scope.go:117] "RemoveContainer" containerID="c88e5cfec1c623e5ad69a5475fcdced12cfc88cec7a26adca79c971e60e94a4c" Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.602583 4745 scope.go:117] "RemoveContainer" containerID="407299999e0652c4be91e60317e1d828fa2df51ebb733566f6f3d178902139f3" Oct 10 13:52:24 crc kubenswrapper[4745]: E1010 13:52:24.603328 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"407299999e0652c4be91e60317e1d828fa2df51ebb733566f6f3d178902139f3\": container with ID starting with 407299999e0652c4be91e60317e1d828fa2df51ebb733566f6f3d178902139f3 not found: ID does not exist" containerID="407299999e0652c4be91e60317e1d828fa2df51ebb733566f6f3d178902139f3" Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.603389 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"407299999e0652c4be91e60317e1d828fa2df51ebb733566f6f3d178902139f3"} err="failed to get container status \"407299999e0652c4be91e60317e1d828fa2df51ebb733566f6f3d178902139f3\": rpc error: code = NotFound desc = could not find container \"407299999e0652c4be91e60317e1d828fa2df51ebb733566f6f3d178902139f3\": container with ID starting with 407299999e0652c4be91e60317e1d828fa2df51ebb733566f6f3d178902139f3 not found: ID does not exist" Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.603418 4745 scope.go:117] "RemoveContainer" containerID="0672f61add8e9f036084c6bfc5ec58214c61b1dfd98cf47ecde878ddf68191ad" Oct 10 13:52:24 crc kubenswrapper[4745]: E1010 13:52:24.604025 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0672f61add8e9f036084c6bfc5ec58214c61b1dfd98cf47ecde878ddf68191ad\": container with ID starting with 0672f61add8e9f036084c6bfc5ec58214c61b1dfd98cf47ecde878ddf68191ad not found: ID does not exist" containerID="0672f61add8e9f036084c6bfc5ec58214c61b1dfd98cf47ecde878ddf68191ad" Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.604065 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0672f61add8e9f036084c6bfc5ec58214c61b1dfd98cf47ecde878ddf68191ad"} err="failed to get container status \"0672f61add8e9f036084c6bfc5ec58214c61b1dfd98cf47ecde878ddf68191ad\": rpc error: code = NotFound desc = could not find container \"0672f61add8e9f036084c6bfc5ec58214c61b1dfd98cf47ecde878ddf68191ad\": container with ID starting with 0672f61add8e9f036084c6bfc5ec58214c61b1dfd98cf47ecde878ddf68191ad not found: ID does not exist" Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.604093 4745 scope.go:117] "RemoveContainer" containerID="c88e5cfec1c623e5ad69a5475fcdced12cfc88cec7a26adca79c971e60e94a4c" Oct 10 13:52:24 crc kubenswrapper[4745]: E1010 13:52:24.604489 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c88e5cfec1c623e5ad69a5475fcdced12cfc88cec7a26adca79c971e60e94a4c\": container with ID starting with c88e5cfec1c623e5ad69a5475fcdced12cfc88cec7a26adca79c971e60e94a4c not found: ID does not exist" containerID="c88e5cfec1c623e5ad69a5475fcdced12cfc88cec7a26adca79c971e60e94a4c" Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.604540 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c88e5cfec1c623e5ad69a5475fcdced12cfc88cec7a26adca79c971e60e94a4c"} err="failed to get container status \"c88e5cfec1c623e5ad69a5475fcdced12cfc88cec7a26adca79c971e60e94a4c\": rpc error: code = NotFound desc = could not find container \"c88e5cfec1c623e5ad69a5475fcdced12cfc88cec7a26adca79c971e60e94a4c\": container with ID starting with c88e5cfec1c623e5ad69a5475fcdced12cfc88cec7a26adca79c971e60e94a4c not found: ID does not exist" Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.776313 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8514ec89-60e8-471b-9476-80f9255695d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8514ec89-60e8-471b-9476-80f9255695d4" (UID: "8514ec89-60e8-471b-9476-80f9255695d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.846072 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gk5tk"] Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.860354 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gk5tk"] Oct 10 13:52:24 crc kubenswrapper[4745]: I1010 13:52:24.862980 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8514ec89-60e8-471b-9476-80f9255695d4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:52:26 crc kubenswrapper[4745]: I1010 13:52:26.766509 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8514ec89-60e8-471b-9476-80f9255695d4" path="/var/lib/kubelet/pods/8514ec89-60e8-471b-9476-80f9255695d4/volumes" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.607492 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w56sd"] Oct 10 13:53:10 crc kubenswrapper[4745]: E1010 13:53:10.608468 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8514ec89-60e8-471b-9476-80f9255695d4" containerName="registry-server" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.608482 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="8514ec89-60e8-471b-9476-80f9255695d4" containerName="registry-server" Oct 10 13:53:10 crc kubenswrapper[4745]: E1010 13:53:10.608500 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8514ec89-60e8-471b-9476-80f9255695d4" containerName="extract-utilities" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.608508 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="8514ec89-60e8-471b-9476-80f9255695d4" containerName="extract-utilities" Oct 10 13:53:10 crc kubenswrapper[4745]: E1010 13:53:10.608521 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8514ec89-60e8-471b-9476-80f9255695d4" containerName="extract-content" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.608527 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="8514ec89-60e8-471b-9476-80f9255695d4" containerName="extract-content" Oct 10 13:53:10 crc kubenswrapper[4745]: E1010 13:53:10.608535 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a95002a-e1dd-42a5-9375-2743704192c1" containerName="extract-content" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.608540 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a95002a-e1dd-42a5-9375-2743704192c1" containerName="extract-content" Oct 10 13:53:10 crc kubenswrapper[4745]: E1010 13:53:10.608563 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a95002a-e1dd-42a5-9375-2743704192c1" containerName="registry-server" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.608569 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a95002a-e1dd-42a5-9375-2743704192c1" containerName="registry-server" Oct 10 13:53:10 crc kubenswrapper[4745]: E1010 13:53:10.608583 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a95002a-e1dd-42a5-9375-2743704192c1" containerName="extract-utilities" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.608588 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a95002a-e1dd-42a5-9375-2743704192c1" containerName="extract-utilities" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.608775 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a95002a-e1dd-42a5-9375-2743704192c1" containerName="registry-server" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.608806 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="8514ec89-60e8-471b-9476-80f9255695d4" containerName="registry-server" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.611059 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w56sd" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.628958 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w56sd"] Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.730068 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb9hl\" (UniqueName: \"kubernetes.io/projected/08d443d6-7407-4048-a3e8-bbd406e2d517-kube-api-access-rb9hl\") pod \"community-operators-w56sd\" (UID: \"08d443d6-7407-4048-a3e8-bbd406e2d517\") " pod="openshift-marketplace/community-operators-w56sd" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.730142 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08d443d6-7407-4048-a3e8-bbd406e2d517-utilities\") pod \"community-operators-w56sd\" (UID: \"08d443d6-7407-4048-a3e8-bbd406e2d517\") " pod="openshift-marketplace/community-operators-w56sd" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.730174 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08d443d6-7407-4048-a3e8-bbd406e2d517-catalog-content\") pod \"community-operators-w56sd\" (UID: \"08d443d6-7407-4048-a3e8-bbd406e2d517\") " pod="openshift-marketplace/community-operators-w56sd" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.831822 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb9hl\" (UniqueName: \"kubernetes.io/projected/08d443d6-7407-4048-a3e8-bbd406e2d517-kube-api-access-rb9hl\") pod \"community-operators-w56sd\" (UID: \"08d443d6-7407-4048-a3e8-bbd406e2d517\") " pod="openshift-marketplace/community-operators-w56sd" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.831964 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08d443d6-7407-4048-a3e8-bbd406e2d517-utilities\") pod \"community-operators-w56sd\" (UID: \"08d443d6-7407-4048-a3e8-bbd406e2d517\") " pod="openshift-marketplace/community-operators-w56sd" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.832069 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08d443d6-7407-4048-a3e8-bbd406e2d517-catalog-content\") pod \"community-operators-w56sd\" (UID: \"08d443d6-7407-4048-a3e8-bbd406e2d517\") " pod="openshift-marketplace/community-operators-w56sd" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.832547 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08d443d6-7407-4048-a3e8-bbd406e2d517-catalog-content\") pod \"community-operators-w56sd\" (UID: \"08d443d6-7407-4048-a3e8-bbd406e2d517\") " pod="openshift-marketplace/community-operators-w56sd" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.833074 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08d443d6-7407-4048-a3e8-bbd406e2d517-utilities\") pod \"community-operators-w56sd\" (UID: \"08d443d6-7407-4048-a3e8-bbd406e2d517\") " pod="openshift-marketplace/community-operators-w56sd" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.855781 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb9hl\" (UniqueName: \"kubernetes.io/projected/08d443d6-7407-4048-a3e8-bbd406e2d517-kube-api-access-rb9hl\") pod \"community-operators-w56sd\" (UID: \"08d443d6-7407-4048-a3e8-bbd406e2d517\") " pod="openshift-marketplace/community-operators-w56sd" Oct 10 13:53:10 crc kubenswrapper[4745]: I1010 13:53:10.979548 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w56sd" Oct 10 13:53:11 crc kubenswrapper[4745]: I1010 13:53:11.482398 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w56sd"] Oct 10 13:53:12 crc kubenswrapper[4745]: I1010 13:53:12.030434 4745 generic.go:334] "Generic (PLEG): container finished" podID="08d443d6-7407-4048-a3e8-bbd406e2d517" containerID="523b2186897f7c41d863c7b0f74c5e29995516bafd46432d9517ec97976e5387" exitCode=0 Oct 10 13:53:12 crc kubenswrapper[4745]: I1010 13:53:12.030576 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w56sd" event={"ID":"08d443d6-7407-4048-a3e8-bbd406e2d517","Type":"ContainerDied","Data":"523b2186897f7c41d863c7b0f74c5e29995516bafd46432d9517ec97976e5387"} Oct 10 13:53:12 crc kubenswrapper[4745]: I1010 13:53:12.030820 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w56sd" event={"ID":"08d443d6-7407-4048-a3e8-bbd406e2d517","Type":"ContainerStarted","Data":"2afff84c815a29d34cf0570d6325c7f5abba820fbae42318fa4b673ae31ef3fd"} Oct 10 13:53:14 crc kubenswrapper[4745]: I1010 13:53:14.056105 4745 generic.go:334] "Generic (PLEG): container finished" podID="08d443d6-7407-4048-a3e8-bbd406e2d517" containerID="6da518627a0104056c50a79e4a83d49ca586d8044d1e98d4c01ca5425877657c" exitCode=0 Oct 10 13:53:14 crc kubenswrapper[4745]: I1010 13:53:14.056309 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w56sd" event={"ID":"08d443d6-7407-4048-a3e8-bbd406e2d517","Type":"ContainerDied","Data":"6da518627a0104056c50a79e4a83d49ca586d8044d1e98d4c01ca5425877657c"} Oct 10 13:53:15 crc kubenswrapper[4745]: I1010 13:53:15.066782 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w56sd" event={"ID":"08d443d6-7407-4048-a3e8-bbd406e2d517","Type":"ContainerStarted","Data":"273ba1aa3b502ed6930bbb1a72c6432a87d89fc4cf10a786352f4fd08a84e117"} Oct 10 13:53:15 crc kubenswrapper[4745]: I1010 13:53:15.099423 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w56sd" podStartSLOduration=2.52582315 podStartE2EDuration="5.099400583s" podCreationTimestamp="2025-10-10 13:53:10 +0000 UTC" firstStartedPulling="2025-10-10 13:53:12.032919488 +0000 UTC m=+2105.930576261" lastFinishedPulling="2025-10-10 13:53:14.606496921 +0000 UTC m=+2108.504153694" observedRunningTime="2025-10-10 13:53:15.087402549 +0000 UTC m=+2108.985059332" watchObservedRunningTime="2025-10-10 13:53:15.099400583 +0000 UTC m=+2108.997057356" Oct 10 13:53:16 crc kubenswrapper[4745]: I1010 13:53:16.187084 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:53:16 crc kubenswrapper[4745]: I1010 13:53:16.187549 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:53:20 crc kubenswrapper[4745]: I1010 13:53:20.996130 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w56sd" Oct 10 13:53:20 crc kubenswrapper[4745]: I1010 13:53:20.997254 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w56sd" Oct 10 13:53:21 crc kubenswrapper[4745]: I1010 13:53:21.039367 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w56sd" Oct 10 13:53:21 crc kubenswrapper[4745]: I1010 13:53:21.194628 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w56sd" Oct 10 13:53:21 crc kubenswrapper[4745]: I1010 13:53:21.294157 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w56sd"] Oct 10 13:53:23 crc kubenswrapper[4745]: I1010 13:53:23.143899 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w56sd" podUID="08d443d6-7407-4048-a3e8-bbd406e2d517" containerName="registry-server" containerID="cri-o://273ba1aa3b502ed6930bbb1a72c6432a87d89fc4cf10a786352f4fd08a84e117" gracePeriod=2 Oct 10 13:53:23 crc kubenswrapper[4745]: I1010 13:53:23.628597 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w56sd" Oct 10 13:53:23 crc kubenswrapper[4745]: I1010 13:53:23.655363 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rb9hl\" (UniqueName: \"kubernetes.io/projected/08d443d6-7407-4048-a3e8-bbd406e2d517-kube-api-access-rb9hl\") pod \"08d443d6-7407-4048-a3e8-bbd406e2d517\" (UID: \"08d443d6-7407-4048-a3e8-bbd406e2d517\") " Oct 10 13:53:23 crc kubenswrapper[4745]: I1010 13:53:23.655651 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08d443d6-7407-4048-a3e8-bbd406e2d517-catalog-content\") pod \"08d443d6-7407-4048-a3e8-bbd406e2d517\" (UID: \"08d443d6-7407-4048-a3e8-bbd406e2d517\") " Oct 10 13:53:23 crc kubenswrapper[4745]: I1010 13:53:23.655702 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08d443d6-7407-4048-a3e8-bbd406e2d517-utilities\") pod \"08d443d6-7407-4048-a3e8-bbd406e2d517\" (UID: \"08d443d6-7407-4048-a3e8-bbd406e2d517\") " Oct 10 13:53:23 crc kubenswrapper[4745]: I1010 13:53:23.660072 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08d443d6-7407-4048-a3e8-bbd406e2d517-utilities" (OuterVolumeSpecName: "utilities") pod "08d443d6-7407-4048-a3e8-bbd406e2d517" (UID: "08d443d6-7407-4048-a3e8-bbd406e2d517"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:53:23 crc kubenswrapper[4745]: I1010 13:53:23.666699 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08d443d6-7407-4048-a3e8-bbd406e2d517-kube-api-access-rb9hl" (OuterVolumeSpecName: "kube-api-access-rb9hl") pod "08d443d6-7407-4048-a3e8-bbd406e2d517" (UID: "08d443d6-7407-4048-a3e8-bbd406e2d517"). InnerVolumeSpecName "kube-api-access-rb9hl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:53:23 crc kubenswrapper[4745]: I1010 13:53:23.708719 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08d443d6-7407-4048-a3e8-bbd406e2d517-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08d443d6-7407-4048-a3e8-bbd406e2d517" (UID: "08d443d6-7407-4048-a3e8-bbd406e2d517"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 13:53:23 crc kubenswrapper[4745]: I1010 13:53:23.757214 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rb9hl\" (UniqueName: \"kubernetes.io/projected/08d443d6-7407-4048-a3e8-bbd406e2d517-kube-api-access-rb9hl\") on node \"crc\" DevicePath \"\"" Oct 10 13:53:23 crc kubenswrapper[4745]: I1010 13:53:23.757242 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08d443d6-7407-4048-a3e8-bbd406e2d517-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 13:53:23 crc kubenswrapper[4745]: I1010 13:53:23.757252 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08d443d6-7407-4048-a3e8-bbd406e2d517-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 13:53:24 crc kubenswrapper[4745]: I1010 13:53:24.157095 4745 generic.go:334] "Generic (PLEG): container finished" podID="08d443d6-7407-4048-a3e8-bbd406e2d517" containerID="273ba1aa3b502ed6930bbb1a72c6432a87d89fc4cf10a786352f4fd08a84e117" exitCode=0 Oct 10 13:53:24 crc kubenswrapper[4745]: I1010 13:53:24.157164 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w56sd" Oct 10 13:53:24 crc kubenswrapper[4745]: I1010 13:53:24.157163 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w56sd" event={"ID":"08d443d6-7407-4048-a3e8-bbd406e2d517","Type":"ContainerDied","Data":"273ba1aa3b502ed6930bbb1a72c6432a87d89fc4cf10a786352f4fd08a84e117"} Oct 10 13:53:24 crc kubenswrapper[4745]: I1010 13:53:24.157651 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w56sd" event={"ID":"08d443d6-7407-4048-a3e8-bbd406e2d517","Type":"ContainerDied","Data":"2afff84c815a29d34cf0570d6325c7f5abba820fbae42318fa4b673ae31ef3fd"} Oct 10 13:53:24 crc kubenswrapper[4745]: I1010 13:53:24.157684 4745 scope.go:117] "RemoveContainer" containerID="273ba1aa3b502ed6930bbb1a72c6432a87d89fc4cf10a786352f4fd08a84e117" Oct 10 13:53:24 crc kubenswrapper[4745]: I1010 13:53:24.192235 4745 scope.go:117] "RemoveContainer" containerID="6da518627a0104056c50a79e4a83d49ca586d8044d1e98d4c01ca5425877657c" Oct 10 13:53:24 crc kubenswrapper[4745]: I1010 13:53:24.209690 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w56sd"] Oct 10 13:53:24 crc kubenswrapper[4745]: I1010 13:53:24.215924 4745 scope.go:117] "RemoveContainer" containerID="523b2186897f7c41d863c7b0f74c5e29995516bafd46432d9517ec97976e5387" Oct 10 13:53:24 crc kubenswrapper[4745]: I1010 13:53:24.217980 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w56sd"] Oct 10 13:53:24 crc kubenswrapper[4745]: I1010 13:53:24.283828 4745 scope.go:117] "RemoveContainer" containerID="273ba1aa3b502ed6930bbb1a72c6432a87d89fc4cf10a786352f4fd08a84e117" Oct 10 13:53:24 crc kubenswrapper[4745]: E1010 13:53:24.284864 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"273ba1aa3b502ed6930bbb1a72c6432a87d89fc4cf10a786352f4fd08a84e117\": container with ID starting with 273ba1aa3b502ed6930bbb1a72c6432a87d89fc4cf10a786352f4fd08a84e117 not found: ID does not exist" containerID="273ba1aa3b502ed6930bbb1a72c6432a87d89fc4cf10a786352f4fd08a84e117" Oct 10 13:53:24 crc kubenswrapper[4745]: I1010 13:53:24.284896 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"273ba1aa3b502ed6930bbb1a72c6432a87d89fc4cf10a786352f4fd08a84e117"} err="failed to get container status \"273ba1aa3b502ed6930bbb1a72c6432a87d89fc4cf10a786352f4fd08a84e117\": rpc error: code = NotFound desc = could not find container \"273ba1aa3b502ed6930bbb1a72c6432a87d89fc4cf10a786352f4fd08a84e117\": container with ID starting with 273ba1aa3b502ed6930bbb1a72c6432a87d89fc4cf10a786352f4fd08a84e117 not found: ID does not exist" Oct 10 13:53:24 crc kubenswrapper[4745]: I1010 13:53:24.284921 4745 scope.go:117] "RemoveContainer" containerID="6da518627a0104056c50a79e4a83d49ca586d8044d1e98d4c01ca5425877657c" Oct 10 13:53:24 crc kubenswrapper[4745]: E1010 13:53:24.285649 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6da518627a0104056c50a79e4a83d49ca586d8044d1e98d4c01ca5425877657c\": container with ID starting with 6da518627a0104056c50a79e4a83d49ca586d8044d1e98d4c01ca5425877657c not found: ID does not exist" containerID="6da518627a0104056c50a79e4a83d49ca586d8044d1e98d4c01ca5425877657c" Oct 10 13:53:24 crc kubenswrapper[4745]: I1010 13:53:24.285702 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6da518627a0104056c50a79e4a83d49ca586d8044d1e98d4c01ca5425877657c"} err="failed to get container status \"6da518627a0104056c50a79e4a83d49ca586d8044d1e98d4c01ca5425877657c\": rpc error: code = NotFound desc = could not find container \"6da518627a0104056c50a79e4a83d49ca586d8044d1e98d4c01ca5425877657c\": container with ID starting with 6da518627a0104056c50a79e4a83d49ca586d8044d1e98d4c01ca5425877657c not found: ID does not exist" Oct 10 13:53:24 crc kubenswrapper[4745]: I1010 13:53:24.285766 4745 scope.go:117] "RemoveContainer" containerID="523b2186897f7c41d863c7b0f74c5e29995516bafd46432d9517ec97976e5387" Oct 10 13:53:24 crc kubenswrapper[4745]: E1010 13:53:24.286275 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"523b2186897f7c41d863c7b0f74c5e29995516bafd46432d9517ec97976e5387\": container with ID starting with 523b2186897f7c41d863c7b0f74c5e29995516bafd46432d9517ec97976e5387 not found: ID does not exist" containerID="523b2186897f7c41d863c7b0f74c5e29995516bafd46432d9517ec97976e5387" Oct 10 13:53:24 crc kubenswrapper[4745]: I1010 13:53:24.286318 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"523b2186897f7c41d863c7b0f74c5e29995516bafd46432d9517ec97976e5387"} err="failed to get container status \"523b2186897f7c41d863c7b0f74c5e29995516bafd46432d9517ec97976e5387\": rpc error: code = NotFound desc = could not find container \"523b2186897f7c41d863c7b0f74c5e29995516bafd46432d9517ec97976e5387\": container with ID starting with 523b2186897f7c41d863c7b0f74c5e29995516bafd46432d9517ec97976e5387 not found: ID does not exist" Oct 10 13:53:24 crc kubenswrapper[4745]: I1010 13:53:24.764493 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08d443d6-7407-4048-a3e8-bbd406e2d517" path="/var/lib/kubelet/pods/08d443d6-7407-4048-a3e8-bbd406e2d517/volumes" Oct 10 13:53:46 crc kubenswrapper[4745]: I1010 13:53:46.186838 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:53:46 crc kubenswrapper[4745]: I1010 13:53:46.187342 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:54:16 crc kubenswrapper[4745]: I1010 13:54:16.187194 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 13:54:16 crc kubenswrapper[4745]: I1010 13:54:16.187751 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 13:54:16 crc kubenswrapper[4745]: I1010 13:54:16.187801 4745 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 13:54:16 crc kubenswrapper[4745]: I1010 13:54:16.188577 4745 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9"} pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 13:54:16 crc kubenswrapper[4745]: I1010 13:54:16.188638 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" containerID="cri-o://631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" gracePeriod=600 Oct 10 13:54:16 crc kubenswrapper[4745]: E1010 13:54:16.331729 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:54:16 crc kubenswrapper[4745]: I1010 13:54:16.770826 4745 generic.go:334] "Generic (PLEG): container finished" podID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" exitCode=0 Oct 10 13:54:16 crc kubenswrapper[4745]: I1010 13:54:16.770896 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerDied","Data":"631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9"} Oct 10 13:54:16 crc kubenswrapper[4745]: I1010 13:54:16.770978 4745 scope.go:117] "RemoveContainer" containerID="ca4a6f66c74821d4fbc4213650a195182b8af7f523a58adff73b04160e45bbe7" Oct 10 13:54:16 crc kubenswrapper[4745]: I1010 13:54:16.771813 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:54:16 crc kubenswrapper[4745]: E1010 13:54:16.772233 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:54:29 crc kubenswrapper[4745]: I1010 13:54:29.744911 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:54:29 crc kubenswrapper[4745]: E1010 13:54:29.745752 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:54:42 crc kubenswrapper[4745]: I1010 13:54:42.746674 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:54:42 crc kubenswrapper[4745]: E1010 13:54:42.747417 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:54:57 crc kubenswrapper[4745]: I1010 13:54:57.745030 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:54:57 crc kubenswrapper[4745]: E1010 13:54:57.745867 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:55:10 crc kubenswrapper[4745]: I1010 13:55:10.745951 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:55:10 crc kubenswrapper[4745]: E1010 13:55:10.746949 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:55:21 crc kubenswrapper[4745]: I1010 13:55:21.745839 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:55:21 crc kubenswrapper[4745]: E1010 13:55:21.747223 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:55:36 crc kubenswrapper[4745]: I1010 13:55:36.752624 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:55:36 crc kubenswrapper[4745]: E1010 13:55:36.753357 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:55:44 crc kubenswrapper[4745]: I1010 13:55:44.665690 4745 generic.go:334] "Generic (PLEG): container finished" podID="4709df73-fa0c-4b05-aa37-8d32a4aa52f1" containerID="08ac879e0a697b88872fc8baf5b65bc74f3bb1959e40a96f1b84cb22be41d537" exitCode=0 Oct 10 13:55:44 crc kubenswrapper[4745]: I1010 13:55:44.665787 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" event={"ID":"4709df73-fa0c-4b05-aa37-8d32a4aa52f1","Type":"ContainerDied","Data":"08ac879e0a697b88872fc8baf5b65bc74f3bb1959e40a96f1b84cb22be41d537"} Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.204528 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.258512 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w7t2\" (UniqueName: \"kubernetes.io/projected/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-kube-api-access-8w7t2\") pod \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.258616 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-inventory\") pod \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.258723 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-libvirt-secret-0\") pod \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.258787 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-ssh-key\") pod \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.258873 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-libvirt-combined-ca-bundle\") pod \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\" (UID: \"4709df73-fa0c-4b05-aa37-8d32a4aa52f1\") " Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.264279 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-kube-api-access-8w7t2" (OuterVolumeSpecName: "kube-api-access-8w7t2") pod "4709df73-fa0c-4b05-aa37-8d32a4aa52f1" (UID: "4709df73-fa0c-4b05-aa37-8d32a4aa52f1"). InnerVolumeSpecName "kube-api-access-8w7t2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.271275 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "4709df73-fa0c-4b05-aa37-8d32a4aa52f1" (UID: "4709df73-fa0c-4b05-aa37-8d32a4aa52f1"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.290094 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-inventory" (OuterVolumeSpecName: "inventory") pod "4709df73-fa0c-4b05-aa37-8d32a4aa52f1" (UID: "4709df73-fa0c-4b05-aa37-8d32a4aa52f1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.297323 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "4709df73-fa0c-4b05-aa37-8d32a4aa52f1" (UID: "4709df73-fa0c-4b05-aa37-8d32a4aa52f1"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.312373 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4709df73-fa0c-4b05-aa37-8d32a4aa52f1" (UID: "4709df73-fa0c-4b05-aa37-8d32a4aa52f1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.361820 4745 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.361856 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.361891 4745 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.361907 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w7t2\" (UniqueName: \"kubernetes.io/projected/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-kube-api-access-8w7t2\") on node \"crc\" DevicePath \"\"" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.361918 4745 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4709df73-fa0c-4b05-aa37-8d32a4aa52f1-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.688821 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" event={"ID":"4709df73-fa0c-4b05-aa37-8d32a4aa52f1","Type":"ContainerDied","Data":"cd596187c6227fbb131960627f3c8c4135c9ad107def65fd46333371771e0692"} Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.688867 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd596187c6227fbb131960627f3c8c4135c9ad107def65fd46333371771e0692" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.688888 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.800723 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s"] Oct 10 13:55:46 crc kubenswrapper[4745]: E1010 13:55:46.801146 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08d443d6-7407-4048-a3e8-bbd406e2d517" containerName="registry-server" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.801166 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="08d443d6-7407-4048-a3e8-bbd406e2d517" containerName="registry-server" Oct 10 13:55:46 crc kubenswrapper[4745]: E1010 13:55:46.801189 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4709df73-fa0c-4b05-aa37-8d32a4aa52f1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.801199 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="4709df73-fa0c-4b05-aa37-8d32a4aa52f1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 10 13:55:46 crc kubenswrapper[4745]: E1010 13:55:46.801214 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08d443d6-7407-4048-a3e8-bbd406e2d517" containerName="extract-utilities" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.801226 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="08d443d6-7407-4048-a3e8-bbd406e2d517" containerName="extract-utilities" Oct 10 13:55:46 crc kubenswrapper[4745]: E1010 13:55:46.801250 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08d443d6-7407-4048-a3e8-bbd406e2d517" containerName="extract-content" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.801258 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="08d443d6-7407-4048-a3e8-bbd406e2d517" containerName="extract-content" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.801481 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="08d443d6-7407-4048-a3e8-bbd406e2d517" containerName="registry-server" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.801497 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="4709df73-fa0c-4b05-aa37-8d32a4aa52f1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.802213 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.804198 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.804680 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.808231 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.808517 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.808586 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.808815 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.809301 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.820913 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s"] Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.870480 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.870653 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.870685 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/274ab6cc-4796-493d-ad35-5cf55195151c-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.870712 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znqlx\" (UniqueName: \"kubernetes.io/projected/274ab6cc-4796-493d-ad35-5cf55195151c-kube-api-access-znqlx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.870775 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.870808 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.870838 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.870895 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.870947 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.973200 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.973272 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/274ab6cc-4796-493d-ad35-5cf55195151c-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.973317 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znqlx\" (UniqueName: \"kubernetes.io/projected/274ab6cc-4796-493d-ad35-5cf55195151c-kube-api-access-znqlx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.973357 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.973408 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.973445 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.973508 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.973568 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.973983 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.974211 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/274ab6cc-4796-493d-ad35-5cf55195151c-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.976948 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.976952 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.977231 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.977454 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.977800 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.978157 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.986560 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:46 crc kubenswrapper[4745]: I1010 13:55:46.991812 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znqlx\" (UniqueName: \"kubernetes.io/projected/274ab6cc-4796-493d-ad35-5cf55195151c-kube-api-access-znqlx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nww5s\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:47 crc kubenswrapper[4745]: I1010 13:55:47.120275 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:55:47 crc kubenswrapper[4745]: I1010 13:55:47.715990 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s"] Oct 10 13:55:47 crc kubenswrapper[4745]: I1010 13:55:47.726325 4745 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 13:55:48 crc kubenswrapper[4745]: I1010 13:55:48.711231 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" event={"ID":"274ab6cc-4796-493d-ad35-5cf55195151c","Type":"ContainerStarted","Data":"0b6813162cdf8e8a2e1024df787c915f3ab7c265cf8f2f283202b5dd30c1f763"} Oct 10 13:55:48 crc kubenswrapper[4745]: I1010 13:55:48.711570 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" event={"ID":"274ab6cc-4796-493d-ad35-5cf55195151c","Type":"ContainerStarted","Data":"1b1680d23aa0a109a9c8f020e59fec77d5c3cab484d5a15acc8895da4d5d534a"} Oct 10 13:55:48 crc kubenswrapper[4745]: I1010 13:55:48.732985 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" podStartSLOduration=2.189702482 podStartE2EDuration="2.732971374s" podCreationTimestamp="2025-10-10 13:55:46 +0000 UTC" firstStartedPulling="2025-10-10 13:55:47.726122244 +0000 UTC m=+2261.623779007" lastFinishedPulling="2025-10-10 13:55:48.269391106 +0000 UTC m=+2262.167047899" observedRunningTime="2025-10-10 13:55:48.729475456 +0000 UTC m=+2262.627132219" watchObservedRunningTime="2025-10-10 13:55:48.732971374 +0000 UTC m=+2262.630628137" Oct 10 13:55:50 crc kubenswrapper[4745]: I1010 13:55:50.746080 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:55:50 crc kubenswrapper[4745]: E1010 13:55:50.746881 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:56:04 crc kubenswrapper[4745]: I1010 13:56:04.744612 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:56:04 crc kubenswrapper[4745]: E1010 13:56:04.745472 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:56:16 crc kubenswrapper[4745]: I1010 13:56:16.751620 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:56:16 crc kubenswrapper[4745]: E1010 13:56:16.752306 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:56:28 crc kubenswrapper[4745]: I1010 13:56:28.745045 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:56:28 crc kubenswrapper[4745]: E1010 13:56:28.746066 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:56:41 crc kubenswrapper[4745]: I1010 13:56:41.746282 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:56:41 crc kubenswrapper[4745]: E1010 13:56:41.749100 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:56:53 crc kubenswrapper[4745]: I1010 13:56:53.744891 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:56:53 crc kubenswrapper[4745]: E1010 13:56:53.745952 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:57:08 crc kubenswrapper[4745]: I1010 13:57:08.744945 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:57:08 crc kubenswrapper[4745]: E1010 13:57:08.745926 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:57:19 crc kubenswrapper[4745]: I1010 13:57:19.745343 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:57:19 crc kubenswrapper[4745]: E1010 13:57:19.746523 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:57:34 crc kubenswrapper[4745]: I1010 13:57:34.745952 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:57:34 crc kubenswrapper[4745]: E1010 13:57:34.747038 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:57:46 crc kubenswrapper[4745]: I1010 13:57:46.764369 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:57:46 crc kubenswrapper[4745]: E1010 13:57:46.765084 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:58:00 crc kubenswrapper[4745]: I1010 13:58:00.745395 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:58:00 crc kubenswrapper[4745]: E1010 13:58:00.746363 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:58:12 crc kubenswrapper[4745]: I1010 13:58:12.744863 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:58:12 crc kubenswrapper[4745]: E1010 13:58:12.745572 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:58:23 crc kubenswrapper[4745]: I1010 13:58:23.745075 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:58:23 crc kubenswrapper[4745]: E1010 13:58:23.745803 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:58:34 crc kubenswrapper[4745]: I1010 13:58:34.745000 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:58:34 crc kubenswrapper[4745]: E1010 13:58:34.746190 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:58:47 crc kubenswrapper[4745]: I1010 13:58:47.745411 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:58:47 crc kubenswrapper[4745]: E1010 13:58:47.746896 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:58:59 crc kubenswrapper[4745]: I1010 13:58:59.745512 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:58:59 crc kubenswrapper[4745]: E1010 13:58:59.746515 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:59:12 crc kubenswrapper[4745]: I1010 13:59:12.745771 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:59:12 crc kubenswrapper[4745]: E1010 13:59:12.747156 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 13:59:21 crc kubenswrapper[4745]: I1010 13:59:21.923363 4745 generic.go:334] "Generic (PLEG): container finished" podID="274ab6cc-4796-493d-ad35-5cf55195151c" containerID="0b6813162cdf8e8a2e1024df787c915f3ab7c265cf8f2f283202b5dd30c1f763" exitCode=0 Oct 10 13:59:21 crc kubenswrapper[4745]: I1010 13:59:21.923579 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" event={"ID":"274ab6cc-4796-493d-ad35-5cf55195151c","Type":"ContainerDied","Data":"0b6813162cdf8e8a2e1024df787c915f3ab7c265cf8f2f283202b5dd30c1f763"} Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.403856 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.563603 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/274ab6cc-4796-493d-ad35-5cf55195151c-nova-extra-config-0\") pod \"274ab6cc-4796-493d-ad35-5cf55195151c\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.564068 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-cell1-compute-config-0\") pod \"274ab6cc-4796-493d-ad35-5cf55195151c\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.564150 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-migration-ssh-key-1\") pod \"274ab6cc-4796-493d-ad35-5cf55195151c\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.564200 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-ssh-key\") pod \"274ab6cc-4796-493d-ad35-5cf55195151c\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.564237 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-inventory\") pod \"274ab6cc-4796-493d-ad35-5cf55195151c\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.564294 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-migration-ssh-key-0\") pod \"274ab6cc-4796-493d-ad35-5cf55195151c\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.564387 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znqlx\" (UniqueName: \"kubernetes.io/projected/274ab6cc-4796-493d-ad35-5cf55195151c-kube-api-access-znqlx\") pod \"274ab6cc-4796-493d-ad35-5cf55195151c\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.564472 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-cell1-compute-config-1\") pod \"274ab6cc-4796-493d-ad35-5cf55195151c\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.564509 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-combined-ca-bundle\") pod \"274ab6cc-4796-493d-ad35-5cf55195151c\" (UID: \"274ab6cc-4796-493d-ad35-5cf55195151c\") " Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.571209 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/274ab6cc-4796-493d-ad35-5cf55195151c-kube-api-access-znqlx" (OuterVolumeSpecName: "kube-api-access-znqlx") pod "274ab6cc-4796-493d-ad35-5cf55195151c" (UID: "274ab6cc-4796-493d-ad35-5cf55195151c"). InnerVolumeSpecName "kube-api-access-znqlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.571341 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "274ab6cc-4796-493d-ad35-5cf55195151c" (UID: "274ab6cc-4796-493d-ad35-5cf55195151c"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.605976 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "274ab6cc-4796-493d-ad35-5cf55195151c" (UID: "274ab6cc-4796-493d-ad35-5cf55195151c"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.612497 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "274ab6cc-4796-493d-ad35-5cf55195151c" (UID: "274ab6cc-4796-493d-ad35-5cf55195151c"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.616396 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "274ab6cc-4796-493d-ad35-5cf55195151c" (UID: "274ab6cc-4796-493d-ad35-5cf55195151c"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.621040 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-inventory" (OuterVolumeSpecName: "inventory") pod "274ab6cc-4796-493d-ad35-5cf55195151c" (UID: "274ab6cc-4796-493d-ad35-5cf55195151c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.630751 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/274ab6cc-4796-493d-ad35-5cf55195151c-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "274ab6cc-4796-493d-ad35-5cf55195151c" (UID: "274ab6cc-4796-493d-ad35-5cf55195151c"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.636887 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "274ab6cc-4796-493d-ad35-5cf55195151c" (UID: "274ab6cc-4796-493d-ad35-5cf55195151c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.644938 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "274ab6cc-4796-493d-ad35-5cf55195151c" (UID: "274ab6cc-4796-493d-ad35-5cf55195151c"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.666359 4745 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.666398 4745 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.666411 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.666423 4745 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.666436 4745 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.666448 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znqlx\" (UniqueName: \"kubernetes.io/projected/274ab6cc-4796-493d-ad35-5cf55195151c-kube-api-access-znqlx\") on node \"crc\" DevicePath \"\"" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.666460 4745 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.666472 4745 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/274ab6cc-4796-493d-ad35-5cf55195151c-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.666508 4745 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/274ab6cc-4796-493d-ad35-5cf55195151c-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.943383 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" event={"ID":"274ab6cc-4796-493d-ad35-5cf55195151c","Type":"ContainerDied","Data":"1b1680d23aa0a109a9c8f020e59fec77d5c3cab484d5a15acc8895da4d5d534a"} Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.943436 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b1680d23aa0a109a9c8f020e59fec77d5c3cab484d5a15acc8895da4d5d534a" Oct 10 13:59:23 crc kubenswrapper[4745]: I1010 13:59:23.943451 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nww5s" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.044555 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7"] Oct 10 13:59:24 crc kubenswrapper[4745]: E1010 13:59:24.044948 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="274ab6cc-4796-493d-ad35-5cf55195151c" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.044965 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="274ab6cc-4796-493d-ad35-5cf55195151c" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.045191 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="274ab6cc-4796-493d-ad35-5cf55195151c" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.046210 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.049797 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.050245 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.051477 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6vb5b" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.051683 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.052594 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.067015 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7"] Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.184958 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.185060 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.185082 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.185123 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2s7l\" (UniqueName: \"kubernetes.io/projected/9c27594e-14e8-4b37-860c-19d4c47605cc-kube-api-access-r2s7l\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.185264 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.185364 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.185673 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.287358 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.287442 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.287536 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.287567 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.287624 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2s7l\" (UniqueName: \"kubernetes.io/projected/9c27594e-14e8-4b37-860c-19d4c47605cc-kube-api-access-r2s7l\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.287652 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.287689 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.292709 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.292784 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.293139 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.293678 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.294382 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.295243 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.305890 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2s7l\" (UniqueName: \"kubernetes.io/projected/9c27594e-14e8-4b37-860c-19d4c47605cc-kube-api-access-r2s7l\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.385193 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 13:59:24 crc kubenswrapper[4745]: I1010 13:59:24.997239 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7"] Oct 10 13:59:25 crc kubenswrapper[4745]: W1010 13:59:25.000481 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c27594e_14e8_4b37_860c_19d4c47605cc.slice/crio-4bb488a0dbcf8f859c4b2a18b156136122c060ec3ee4d369b875796045b53b48 WatchSource:0}: Error finding container 4bb488a0dbcf8f859c4b2a18b156136122c060ec3ee4d369b875796045b53b48: Status 404 returned error can't find the container with id 4bb488a0dbcf8f859c4b2a18b156136122c060ec3ee4d369b875796045b53b48 Oct 10 13:59:25 crc kubenswrapper[4745]: I1010 13:59:25.744798 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 13:59:25 crc kubenswrapper[4745]: I1010 13:59:25.965047 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" event={"ID":"9c27594e-14e8-4b37-860c-19d4c47605cc","Type":"ContainerStarted","Data":"4bb488a0dbcf8f859c4b2a18b156136122c060ec3ee4d369b875796045b53b48"} Oct 10 13:59:26 crc kubenswrapper[4745]: I1010 13:59:26.975466 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"1f503d509545908a6f3ef6fd56dd209acab0d1ff66704707b763337227936e6d"} Oct 10 13:59:26 crc kubenswrapper[4745]: I1010 13:59:26.977070 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" event={"ID":"9c27594e-14e8-4b37-860c-19d4c47605cc","Type":"ContainerStarted","Data":"5cfe6d1789beeb6d23bc5aafbb789cd63a0d75b69628662379ea297cee913357"} Oct 10 13:59:27 crc kubenswrapper[4745]: I1010 13:59:27.018531 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" podStartSLOduration=2.39388572 podStartE2EDuration="3.018506123s" podCreationTimestamp="2025-10-10 13:59:24 +0000 UTC" firstStartedPulling="2025-10-10 13:59:25.003810315 +0000 UTC m=+2478.901467078" lastFinishedPulling="2025-10-10 13:59:25.628430688 +0000 UTC m=+2479.526087481" observedRunningTime="2025-10-10 13:59:27.009839098 +0000 UTC m=+2480.907495861" watchObservedRunningTime="2025-10-10 13:59:27.018506123 +0000 UTC m=+2480.916162896" Oct 10 14:00:00 crc kubenswrapper[4745]: I1010 14:00:00.147140 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh"] Oct 10 14:00:00 crc kubenswrapper[4745]: I1010 14:00:00.148901 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh" Oct 10 14:00:00 crc kubenswrapper[4745]: I1010 14:00:00.151525 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 14:00:00 crc kubenswrapper[4745]: I1010 14:00:00.152017 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 14:00:00 crc kubenswrapper[4745]: I1010 14:00:00.164216 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh"] Oct 10 14:00:00 crc kubenswrapper[4745]: I1010 14:00:00.273874 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-secret-volume\") pod \"collect-profiles-29335080-fmphh\" (UID: \"b3dc410d-d0a0-4b39-9d02-78271e07fe3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh" Oct 10 14:00:00 crc kubenswrapper[4745]: I1010 14:00:00.274132 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmtlw\" (UniqueName: \"kubernetes.io/projected/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-kube-api-access-bmtlw\") pod \"collect-profiles-29335080-fmphh\" (UID: \"b3dc410d-d0a0-4b39-9d02-78271e07fe3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh" Oct 10 14:00:00 crc kubenswrapper[4745]: I1010 14:00:00.274171 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-config-volume\") pod \"collect-profiles-29335080-fmphh\" (UID: \"b3dc410d-d0a0-4b39-9d02-78271e07fe3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh" Oct 10 14:00:00 crc kubenswrapper[4745]: I1010 14:00:00.382248 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmtlw\" (UniqueName: \"kubernetes.io/projected/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-kube-api-access-bmtlw\") pod \"collect-profiles-29335080-fmphh\" (UID: \"b3dc410d-d0a0-4b39-9d02-78271e07fe3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh" Oct 10 14:00:00 crc kubenswrapper[4745]: I1010 14:00:00.382321 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-config-volume\") pod \"collect-profiles-29335080-fmphh\" (UID: \"b3dc410d-d0a0-4b39-9d02-78271e07fe3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh" Oct 10 14:00:00 crc kubenswrapper[4745]: I1010 14:00:00.382455 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-secret-volume\") pod \"collect-profiles-29335080-fmphh\" (UID: \"b3dc410d-d0a0-4b39-9d02-78271e07fe3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh" Oct 10 14:00:00 crc kubenswrapper[4745]: I1010 14:00:00.383701 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-config-volume\") pod \"collect-profiles-29335080-fmphh\" (UID: \"b3dc410d-d0a0-4b39-9d02-78271e07fe3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh" Oct 10 14:00:00 crc kubenswrapper[4745]: I1010 14:00:00.388092 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-secret-volume\") pod \"collect-profiles-29335080-fmphh\" (UID: \"b3dc410d-d0a0-4b39-9d02-78271e07fe3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh" Oct 10 14:00:00 crc kubenswrapper[4745]: I1010 14:00:00.397954 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmtlw\" (UniqueName: \"kubernetes.io/projected/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-kube-api-access-bmtlw\") pod \"collect-profiles-29335080-fmphh\" (UID: \"b3dc410d-d0a0-4b39-9d02-78271e07fe3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh" Oct 10 14:00:00 crc kubenswrapper[4745]: I1010 14:00:00.474687 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh" Oct 10 14:00:00 crc kubenswrapper[4745]: I1010 14:00:00.904956 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh"] Oct 10 14:00:01 crc kubenswrapper[4745]: I1010 14:00:01.394561 4745 generic.go:334] "Generic (PLEG): container finished" podID="b3dc410d-d0a0-4b39-9d02-78271e07fe3e" containerID="61e3f361ffd17c6283ec75493f649ed3b29664e29b8d86ecc750d405650580b9" exitCode=0 Oct 10 14:00:01 crc kubenswrapper[4745]: I1010 14:00:01.394661 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh" event={"ID":"b3dc410d-d0a0-4b39-9d02-78271e07fe3e","Type":"ContainerDied","Data":"61e3f361ffd17c6283ec75493f649ed3b29664e29b8d86ecc750d405650580b9"} Oct 10 14:00:01 crc kubenswrapper[4745]: I1010 14:00:01.394890 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh" event={"ID":"b3dc410d-d0a0-4b39-9d02-78271e07fe3e","Type":"ContainerStarted","Data":"99499b67a1cc59a63a6b58ceb2c54c2b683bc5db18437ecb9117704acc9e9259"} Oct 10 14:00:02 crc kubenswrapper[4745]: I1010 14:00:02.803266 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh" Oct 10 14:00:02 crc kubenswrapper[4745]: I1010 14:00:02.937783 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-secret-volume\") pod \"b3dc410d-d0a0-4b39-9d02-78271e07fe3e\" (UID: \"b3dc410d-d0a0-4b39-9d02-78271e07fe3e\") " Oct 10 14:00:02 crc kubenswrapper[4745]: I1010 14:00:02.937960 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-config-volume\") pod \"b3dc410d-d0a0-4b39-9d02-78271e07fe3e\" (UID: \"b3dc410d-d0a0-4b39-9d02-78271e07fe3e\") " Oct 10 14:00:02 crc kubenswrapper[4745]: I1010 14:00:02.938004 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmtlw\" (UniqueName: \"kubernetes.io/projected/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-kube-api-access-bmtlw\") pod \"b3dc410d-d0a0-4b39-9d02-78271e07fe3e\" (UID: \"b3dc410d-d0a0-4b39-9d02-78271e07fe3e\") " Oct 10 14:00:02 crc kubenswrapper[4745]: I1010 14:00:02.940137 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-config-volume" (OuterVolumeSpecName: "config-volume") pod "b3dc410d-d0a0-4b39-9d02-78271e07fe3e" (UID: "b3dc410d-d0a0-4b39-9d02-78271e07fe3e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:00:02 crc kubenswrapper[4745]: I1010 14:00:02.944279 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b3dc410d-d0a0-4b39-9d02-78271e07fe3e" (UID: "b3dc410d-d0a0-4b39-9d02-78271e07fe3e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:00:02 crc kubenswrapper[4745]: I1010 14:00:02.947857 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-kube-api-access-bmtlw" (OuterVolumeSpecName: "kube-api-access-bmtlw") pod "b3dc410d-d0a0-4b39-9d02-78271e07fe3e" (UID: "b3dc410d-d0a0-4b39-9d02-78271e07fe3e"). InnerVolumeSpecName "kube-api-access-bmtlw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:00:03 crc kubenswrapper[4745]: I1010 14:00:03.040458 4745 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 14:00:03 crc kubenswrapper[4745]: I1010 14:00:03.040484 4745 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 14:00:03 crc kubenswrapper[4745]: I1010 14:00:03.040493 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmtlw\" (UniqueName: \"kubernetes.io/projected/b3dc410d-d0a0-4b39-9d02-78271e07fe3e-kube-api-access-bmtlw\") on node \"crc\" DevicePath \"\"" Oct 10 14:00:03 crc kubenswrapper[4745]: I1010 14:00:03.417248 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh" event={"ID":"b3dc410d-d0a0-4b39-9d02-78271e07fe3e","Type":"ContainerDied","Data":"99499b67a1cc59a63a6b58ceb2c54c2b683bc5db18437ecb9117704acc9e9259"} Oct 10 14:00:03 crc kubenswrapper[4745]: I1010 14:00:03.417292 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99499b67a1cc59a63a6b58ceb2c54c2b683bc5db18437ecb9117704acc9e9259" Oct 10 14:00:03 crc kubenswrapper[4745]: I1010 14:00:03.417338 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335080-fmphh" Oct 10 14:00:03 crc kubenswrapper[4745]: I1010 14:00:03.903711 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w"] Oct 10 14:00:03 crc kubenswrapper[4745]: I1010 14:00:03.919155 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335035-mtp5w"] Oct 10 14:00:04 crc kubenswrapper[4745]: I1010 14:00:04.759446 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cb8607d-1cbd-4b3a-9b06-f2db94952d82" path="/var/lib/kubelet/pods/5cb8607d-1cbd-4b3a-9b06-f2db94952d82/volumes" Oct 10 14:00:15 crc kubenswrapper[4745]: I1010 14:00:15.039022 4745 scope.go:117] "RemoveContainer" containerID="2ac9fd05a9335664c489379f6352d1d2002c1b227651cf5ebc79284e99e1aca3" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.165064 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29335081-sq6xb"] Oct 10 14:01:00 crc kubenswrapper[4745]: E1010 14:01:00.166539 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3dc410d-d0a0-4b39-9d02-78271e07fe3e" containerName="collect-profiles" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.166575 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3dc410d-d0a0-4b39-9d02-78271e07fe3e" containerName="collect-profiles" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.169214 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3dc410d-d0a0-4b39-9d02-78271e07fe3e" containerName="collect-profiles" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.170654 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29335081-sq6xb" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.188658 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29335081-sq6xb"] Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.273612 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-config-data\") pod \"keystone-cron-29335081-sq6xb\" (UID: \"a5f9c2a5-82c5-456e-b023-667c13b4d328\") " pod="openstack/keystone-cron-29335081-sq6xb" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.273670 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g59sf\" (UniqueName: \"kubernetes.io/projected/a5f9c2a5-82c5-456e-b023-667c13b4d328-kube-api-access-g59sf\") pod \"keystone-cron-29335081-sq6xb\" (UID: \"a5f9c2a5-82c5-456e-b023-667c13b4d328\") " pod="openstack/keystone-cron-29335081-sq6xb" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.273813 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-fernet-keys\") pod \"keystone-cron-29335081-sq6xb\" (UID: \"a5f9c2a5-82c5-456e-b023-667c13b4d328\") " pod="openstack/keystone-cron-29335081-sq6xb" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.273842 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-combined-ca-bundle\") pod \"keystone-cron-29335081-sq6xb\" (UID: \"a5f9c2a5-82c5-456e-b023-667c13b4d328\") " pod="openstack/keystone-cron-29335081-sq6xb" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.375653 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-combined-ca-bundle\") pod \"keystone-cron-29335081-sq6xb\" (UID: \"a5f9c2a5-82c5-456e-b023-667c13b4d328\") " pod="openstack/keystone-cron-29335081-sq6xb" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.375900 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-config-data\") pod \"keystone-cron-29335081-sq6xb\" (UID: \"a5f9c2a5-82c5-456e-b023-667c13b4d328\") " pod="openstack/keystone-cron-29335081-sq6xb" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.375977 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g59sf\" (UniqueName: \"kubernetes.io/projected/a5f9c2a5-82c5-456e-b023-667c13b4d328-kube-api-access-g59sf\") pod \"keystone-cron-29335081-sq6xb\" (UID: \"a5f9c2a5-82c5-456e-b023-667c13b4d328\") " pod="openstack/keystone-cron-29335081-sq6xb" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.376137 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-fernet-keys\") pod \"keystone-cron-29335081-sq6xb\" (UID: \"a5f9c2a5-82c5-456e-b023-667c13b4d328\") " pod="openstack/keystone-cron-29335081-sq6xb" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.382341 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-config-data\") pod \"keystone-cron-29335081-sq6xb\" (UID: \"a5f9c2a5-82c5-456e-b023-667c13b4d328\") " pod="openstack/keystone-cron-29335081-sq6xb" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.383041 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-combined-ca-bundle\") pod \"keystone-cron-29335081-sq6xb\" (UID: \"a5f9c2a5-82c5-456e-b023-667c13b4d328\") " pod="openstack/keystone-cron-29335081-sq6xb" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.387545 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-fernet-keys\") pod \"keystone-cron-29335081-sq6xb\" (UID: \"a5f9c2a5-82c5-456e-b023-667c13b4d328\") " pod="openstack/keystone-cron-29335081-sq6xb" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.397686 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g59sf\" (UniqueName: \"kubernetes.io/projected/a5f9c2a5-82c5-456e-b023-667c13b4d328-kube-api-access-g59sf\") pod \"keystone-cron-29335081-sq6xb\" (UID: \"a5f9c2a5-82c5-456e-b023-667c13b4d328\") " pod="openstack/keystone-cron-29335081-sq6xb" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.495418 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29335081-sq6xb" Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.934534 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29335081-sq6xb"] Oct 10 14:01:00 crc kubenswrapper[4745]: I1010 14:01:00.983443 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29335081-sq6xb" event={"ID":"a5f9c2a5-82c5-456e-b023-667c13b4d328","Type":"ContainerStarted","Data":"454ba597a543325ebfca486203edee191c7b64c5bfc2028dcfec3de706c19d97"} Oct 10 14:01:01 crc kubenswrapper[4745]: I1010 14:01:01.995599 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29335081-sq6xb" event={"ID":"a5f9c2a5-82c5-456e-b023-667c13b4d328","Type":"ContainerStarted","Data":"11660e91d07169532ad0ec956d4cacf0d9c28dea1645f5e541ae8440b2b7d77d"} Oct 10 14:01:02 crc kubenswrapper[4745]: I1010 14:01:02.022340 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29335081-sq6xb" podStartSLOduration=2.022318776 podStartE2EDuration="2.022318776s" podCreationTimestamp="2025-10-10 14:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:01:02.01566353 +0000 UTC m=+2575.913320323" watchObservedRunningTime="2025-10-10 14:01:02.022318776 +0000 UTC m=+2575.919975559" Oct 10 14:01:04 crc kubenswrapper[4745]: I1010 14:01:04.015873 4745 generic.go:334] "Generic (PLEG): container finished" podID="a5f9c2a5-82c5-456e-b023-667c13b4d328" containerID="11660e91d07169532ad0ec956d4cacf0d9c28dea1645f5e541ae8440b2b7d77d" exitCode=0 Oct 10 14:01:04 crc kubenswrapper[4745]: I1010 14:01:04.015923 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29335081-sq6xb" event={"ID":"a5f9c2a5-82c5-456e-b023-667c13b4d328","Type":"ContainerDied","Data":"11660e91d07169532ad0ec956d4cacf0d9c28dea1645f5e541ae8440b2b7d77d"} Oct 10 14:01:05 crc kubenswrapper[4745]: I1010 14:01:05.352666 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29335081-sq6xb" Oct 10 14:01:05 crc kubenswrapper[4745]: I1010 14:01:05.475922 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g59sf\" (UniqueName: \"kubernetes.io/projected/a5f9c2a5-82c5-456e-b023-667c13b4d328-kube-api-access-g59sf\") pod \"a5f9c2a5-82c5-456e-b023-667c13b4d328\" (UID: \"a5f9c2a5-82c5-456e-b023-667c13b4d328\") " Oct 10 14:01:05 crc kubenswrapper[4745]: I1010 14:01:05.475979 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-config-data\") pod \"a5f9c2a5-82c5-456e-b023-667c13b4d328\" (UID: \"a5f9c2a5-82c5-456e-b023-667c13b4d328\") " Oct 10 14:01:05 crc kubenswrapper[4745]: I1010 14:01:05.476141 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-fernet-keys\") pod \"a5f9c2a5-82c5-456e-b023-667c13b4d328\" (UID: \"a5f9c2a5-82c5-456e-b023-667c13b4d328\") " Oct 10 14:01:05 crc kubenswrapper[4745]: I1010 14:01:05.476199 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-combined-ca-bundle\") pod \"a5f9c2a5-82c5-456e-b023-667c13b4d328\" (UID: \"a5f9c2a5-82c5-456e-b023-667c13b4d328\") " Oct 10 14:01:05 crc kubenswrapper[4745]: I1010 14:01:05.481056 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5f9c2a5-82c5-456e-b023-667c13b4d328-kube-api-access-g59sf" (OuterVolumeSpecName: "kube-api-access-g59sf") pod "a5f9c2a5-82c5-456e-b023-667c13b4d328" (UID: "a5f9c2a5-82c5-456e-b023-667c13b4d328"). InnerVolumeSpecName "kube-api-access-g59sf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:01:05 crc kubenswrapper[4745]: I1010 14:01:05.481358 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a5f9c2a5-82c5-456e-b023-667c13b4d328" (UID: "a5f9c2a5-82c5-456e-b023-667c13b4d328"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:01:05 crc kubenswrapper[4745]: I1010 14:01:05.504193 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5f9c2a5-82c5-456e-b023-667c13b4d328" (UID: "a5f9c2a5-82c5-456e-b023-667c13b4d328"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:01:05 crc kubenswrapper[4745]: I1010 14:01:05.537317 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-config-data" (OuterVolumeSpecName: "config-data") pod "a5f9c2a5-82c5-456e-b023-667c13b4d328" (UID: "a5f9c2a5-82c5-456e-b023-667c13b4d328"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:01:05 crc kubenswrapper[4745]: I1010 14:01:05.578390 4745 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 10 14:01:05 crc kubenswrapper[4745]: I1010 14:01:05.578440 4745 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 14:01:05 crc kubenswrapper[4745]: I1010 14:01:05.578459 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g59sf\" (UniqueName: \"kubernetes.io/projected/a5f9c2a5-82c5-456e-b023-667c13b4d328-kube-api-access-g59sf\") on node \"crc\" DevicePath \"\"" Oct 10 14:01:05 crc kubenswrapper[4745]: I1010 14:01:05.578476 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f9c2a5-82c5-456e-b023-667c13b4d328-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 14:01:06 crc kubenswrapper[4745]: I1010 14:01:06.037455 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29335081-sq6xb" event={"ID":"a5f9c2a5-82c5-456e-b023-667c13b4d328","Type":"ContainerDied","Data":"454ba597a543325ebfca486203edee191c7b64c5bfc2028dcfec3de706c19d97"} Oct 10 14:01:06 crc kubenswrapper[4745]: I1010 14:01:06.038104 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="454ba597a543325ebfca486203edee191c7b64c5bfc2028dcfec3de706c19d97" Oct 10 14:01:06 crc kubenswrapper[4745]: I1010 14:01:06.037509 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29335081-sq6xb" Oct 10 14:01:46 crc kubenswrapper[4745]: I1010 14:01:46.186903 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:01:46 crc kubenswrapper[4745]: I1010 14:01:46.187802 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:01:59 crc kubenswrapper[4745]: I1010 14:01:59.634372 4745 generic.go:334] "Generic (PLEG): container finished" podID="9c27594e-14e8-4b37-860c-19d4c47605cc" containerID="5cfe6d1789beeb6d23bc5aafbb789cd63a0d75b69628662379ea297cee913357" exitCode=0 Oct 10 14:01:59 crc kubenswrapper[4745]: I1010 14:01:59.634446 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" event={"ID":"9c27594e-14e8-4b37-860c-19d4c47605cc","Type":"ContainerDied","Data":"5cfe6d1789beeb6d23bc5aafbb789cd63a0d75b69628662379ea297cee913357"} Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.064031 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.177166 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2s7l\" (UniqueName: \"kubernetes.io/projected/9c27594e-14e8-4b37-860c-19d4c47605cc-kube-api-access-r2s7l\") pod \"9c27594e-14e8-4b37-860c-19d4c47605cc\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.177350 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-1\") pod \"9c27594e-14e8-4b37-860c-19d4c47605cc\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.177464 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-telemetry-combined-ca-bundle\") pod \"9c27594e-14e8-4b37-860c-19d4c47605cc\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.178563 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ssh-key\") pod \"9c27594e-14e8-4b37-860c-19d4c47605cc\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.178679 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-inventory\") pod \"9c27594e-14e8-4b37-860c-19d4c47605cc\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.178785 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-2\") pod \"9c27594e-14e8-4b37-860c-19d4c47605cc\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.178827 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-0\") pod \"9c27594e-14e8-4b37-860c-19d4c47605cc\" (UID: \"9c27594e-14e8-4b37-860c-19d4c47605cc\") " Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.183254 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "9c27594e-14e8-4b37-860c-19d4c47605cc" (UID: "9c27594e-14e8-4b37-860c-19d4c47605cc"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.183316 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c27594e-14e8-4b37-860c-19d4c47605cc-kube-api-access-r2s7l" (OuterVolumeSpecName: "kube-api-access-r2s7l") pod "9c27594e-14e8-4b37-860c-19d4c47605cc" (UID: "9c27594e-14e8-4b37-860c-19d4c47605cc"). InnerVolumeSpecName "kube-api-access-r2s7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.211057 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "9c27594e-14e8-4b37-860c-19d4c47605cc" (UID: "9c27594e-14e8-4b37-860c-19d4c47605cc"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.213496 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9c27594e-14e8-4b37-860c-19d4c47605cc" (UID: "9c27594e-14e8-4b37-860c-19d4c47605cc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.229138 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-inventory" (OuterVolumeSpecName: "inventory") pod "9c27594e-14e8-4b37-860c-19d4c47605cc" (UID: "9c27594e-14e8-4b37-860c-19d4c47605cc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.234406 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "9c27594e-14e8-4b37-860c-19d4c47605cc" (UID: "9c27594e-14e8-4b37-860c-19d4c47605cc"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.238145 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "9c27594e-14e8-4b37-860c-19d4c47605cc" (UID: "9c27594e-14e8-4b37-860c-19d4c47605cc"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.283017 4745 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.283067 4745 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.283091 4745 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.283113 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2s7l\" (UniqueName: \"kubernetes.io/projected/9c27594e-14e8-4b37-860c-19d4c47605cc-kube-api-access-r2s7l\") on node \"crc\" DevicePath \"\"" Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.283132 4745 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.283150 4745 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.283168 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c27594e-14e8-4b37-860c-19d4c47605cc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.659674 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" event={"ID":"9c27594e-14e8-4b37-860c-19d4c47605cc","Type":"ContainerDied","Data":"4bb488a0dbcf8f859c4b2a18b156136122c060ec3ee4d369b875796045b53b48"} Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.659709 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bb488a0dbcf8f859c4b2a18b156136122c060ec3ee4d369b875796045b53b48" Oct 10 14:02:01 crc kubenswrapper[4745]: I1010 14:02:01.659765 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7" Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.096422 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qcgxk"] Oct 10 14:02:07 crc kubenswrapper[4745]: E1010 14:02:07.098347 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c27594e-14e8-4b37-860c-19d4c47605cc" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.098390 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c27594e-14e8-4b37-860c-19d4c47605cc" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 10 14:02:07 crc kubenswrapper[4745]: E1010 14:02:07.098442 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5f9c2a5-82c5-456e-b023-667c13b4d328" containerName="keystone-cron" Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.098460 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5f9c2a5-82c5-456e-b023-667c13b4d328" containerName="keystone-cron" Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.099030 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5f9c2a5-82c5-456e-b023-667c13b4d328" containerName="keystone-cron" Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.099085 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c27594e-14e8-4b37-860c-19d4c47605cc" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.102675 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcgxk" Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.125558 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qcgxk"] Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.199180 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6vcx\" (UniqueName: \"kubernetes.io/projected/285bc4f3-b69e-4c69-99a3-72eae3e41980-kube-api-access-q6vcx\") pod \"redhat-operators-qcgxk\" (UID: \"285bc4f3-b69e-4c69-99a3-72eae3e41980\") " pod="openshift-marketplace/redhat-operators-qcgxk" Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.199649 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/285bc4f3-b69e-4c69-99a3-72eae3e41980-catalog-content\") pod \"redhat-operators-qcgxk\" (UID: \"285bc4f3-b69e-4c69-99a3-72eae3e41980\") " pod="openshift-marketplace/redhat-operators-qcgxk" Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.199812 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/285bc4f3-b69e-4c69-99a3-72eae3e41980-utilities\") pod \"redhat-operators-qcgxk\" (UID: \"285bc4f3-b69e-4c69-99a3-72eae3e41980\") " pod="openshift-marketplace/redhat-operators-qcgxk" Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.301177 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6vcx\" (UniqueName: \"kubernetes.io/projected/285bc4f3-b69e-4c69-99a3-72eae3e41980-kube-api-access-q6vcx\") pod \"redhat-operators-qcgxk\" (UID: \"285bc4f3-b69e-4c69-99a3-72eae3e41980\") " pod="openshift-marketplace/redhat-operators-qcgxk" Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.301249 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/285bc4f3-b69e-4c69-99a3-72eae3e41980-catalog-content\") pod \"redhat-operators-qcgxk\" (UID: \"285bc4f3-b69e-4c69-99a3-72eae3e41980\") " pod="openshift-marketplace/redhat-operators-qcgxk" Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.301346 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/285bc4f3-b69e-4c69-99a3-72eae3e41980-utilities\") pod \"redhat-operators-qcgxk\" (UID: \"285bc4f3-b69e-4c69-99a3-72eae3e41980\") " pod="openshift-marketplace/redhat-operators-qcgxk" Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.302055 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/285bc4f3-b69e-4c69-99a3-72eae3e41980-catalog-content\") pod \"redhat-operators-qcgxk\" (UID: \"285bc4f3-b69e-4c69-99a3-72eae3e41980\") " pod="openshift-marketplace/redhat-operators-qcgxk" Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.302077 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/285bc4f3-b69e-4c69-99a3-72eae3e41980-utilities\") pod \"redhat-operators-qcgxk\" (UID: \"285bc4f3-b69e-4c69-99a3-72eae3e41980\") " pod="openshift-marketplace/redhat-operators-qcgxk" Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.325628 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6vcx\" (UniqueName: \"kubernetes.io/projected/285bc4f3-b69e-4c69-99a3-72eae3e41980-kube-api-access-q6vcx\") pod \"redhat-operators-qcgxk\" (UID: \"285bc4f3-b69e-4c69-99a3-72eae3e41980\") " pod="openshift-marketplace/redhat-operators-qcgxk" Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.443134 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcgxk" Oct 10 14:02:07 crc kubenswrapper[4745]: I1010 14:02:07.900008 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qcgxk"] Oct 10 14:02:07 crc kubenswrapper[4745]: W1010 14:02:07.904981 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod285bc4f3_b69e_4c69_99a3_72eae3e41980.slice/crio-59327154f2bd4db7f41614698ab6d479aca568360ba840cb874aa13a1585300f WatchSource:0}: Error finding container 59327154f2bd4db7f41614698ab6d479aca568360ba840cb874aa13a1585300f: Status 404 returned error can't find the container with id 59327154f2bd4db7f41614698ab6d479aca568360ba840cb874aa13a1585300f Oct 10 14:02:08 crc kubenswrapper[4745]: I1010 14:02:08.720043 4745 generic.go:334] "Generic (PLEG): container finished" podID="285bc4f3-b69e-4c69-99a3-72eae3e41980" containerID="c351e02f7c284aa20097380119bfda5f963cd368824be1fbd4e50f4efed6dc97" exitCode=0 Oct 10 14:02:08 crc kubenswrapper[4745]: I1010 14:02:08.720421 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcgxk" event={"ID":"285bc4f3-b69e-4c69-99a3-72eae3e41980","Type":"ContainerDied","Data":"c351e02f7c284aa20097380119bfda5f963cd368824be1fbd4e50f4efed6dc97"} Oct 10 14:02:08 crc kubenswrapper[4745]: I1010 14:02:08.720454 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcgxk" event={"ID":"285bc4f3-b69e-4c69-99a3-72eae3e41980","Type":"ContainerStarted","Data":"59327154f2bd4db7f41614698ab6d479aca568360ba840cb874aa13a1585300f"} Oct 10 14:02:08 crc kubenswrapper[4745]: I1010 14:02:08.722702 4745 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 14:02:09 crc kubenswrapper[4745]: I1010 14:02:09.736026 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcgxk" event={"ID":"285bc4f3-b69e-4c69-99a3-72eae3e41980","Type":"ContainerStarted","Data":"ab234fc267ff14c60024058124085b42b46306266a4b1006b4ee33d0c1c681e6"} Oct 10 14:02:10 crc kubenswrapper[4745]: I1010 14:02:10.746445 4745 generic.go:334] "Generic (PLEG): container finished" podID="285bc4f3-b69e-4c69-99a3-72eae3e41980" containerID="ab234fc267ff14c60024058124085b42b46306266a4b1006b4ee33d0c1c681e6" exitCode=0 Oct 10 14:02:10 crc kubenswrapper[4745]: I1010 14:02:10.756650 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcgxk" event={"ID":"285bc4f3-b69e-4c69-99a3-72eae3e41980","Type":"ContainerDied","Data":"ab234fc267ff14c60024058124085b42b46306266a4b1006b4ee33d0c1c681e6"} Oct 10 14:02:11 crc kubenswrapper[4745]: I1010 14:02:11.758312 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcgxk" event={"ID":"285bc4f3-b69e-4c69-99a3-72eae3e41980","Type":"ContainerStarted","Data":"4388794e716a888de4a38d70eba5d80004b585de8d5c1457953240fd1e09ea33"} Oct 10 14:02:11 crc kubenswrapper[4745]: I1010 14:02:11.781212 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qcgxk" podStartSLOduration=2.152756952 podStartE2EDuration="4.781196142s" podCreationTimestamp="2025-10-10 14:02:07 +0000 UTC" firstStartedPulling="2025-10-10 14:02:08.722415128 +0000 UTC m=+2642.620071891" lastFinishedPulling="2025-10-10 14:02:11.350854318 +0000 UTC m=+2645.248511081" observedRunningTime="2025-10-10 14:02:11.77424406 +0000 UTC m=+2645.671900823" watchObservedRunningTime="2025-10-10 14:02:11.781196142 +0000 UTC m=+2645.678852905" Oct 10 14:02:16 crc kubenswrapper[4745]: I1010 14:02:16.186322 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:02:16 crc kubenswrapper[4745]: I1010 14:02:16.186949 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:02:17 crc kubenswrapper[4745]: I1010 14:02:17.443953 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qcgxk" Oct 10 14:02:17 crc kubenswrapper[4745]: I1010 14:02:17.445019 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qcgxk" Oct 10 14:02:17 crc kubenswrapper[4745]: I1010 14:02:17.503267 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qcgxk" Oct 10 14:02:17 crc kubenswrapper[4745]: I1010 14:02:17.852579 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qcgxk" Oct 10 14:02:17 crc kubenswrapper[4745]: I1010 14:02:17.902972 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qcgxk"] Oct 10 14:02:19 crc kubenswrapper[4745]: I1010 14:02:19.824759 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qcgxk" podUID="285bc4f3-b69e-4c69-99a3-72eae3e41980" containerName="registry-server" containerID="cri-o://4388794e716a888de4a38d70eba5d80004b585de8d5c1457953240fd1e09ea33" gracePeriod=2 Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.265342 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcgxk" Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.457431 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/285bc4f3-b69e-4c69-99a3-72eae3e41980-catalog-content\") pod \"285bc4f3-b69e-4c69-99a3-72eae3e41980\" (UID: \"285bc4f3-b69e-4c69-99a3-72eae3e41980\") " Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.457486 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6vcx\" (UniqueName: \"kubernetes.io/projected/285bc4f3-b69e-4c69-99a3-72eae3e41980-kube-api-access-q6vcx\") pod \"285bc4f3-b69e-4c69-99a3-72eae3e41980\" (UID: \"285bc4f3-b69e-4c69-99a3-72eae3e41980\") " Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.457512 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/285bc4f3-b69e-4c69-99a3-72eae3e41980-utilities\") pod \"285bc4f3-b69e-4c69-99a3-72eae3e41980\" (UID: \"285bc4f3-b69e-4c69-99a3-72eae3e41980\") " Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.458737 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/285bc4f3-b69e-4c69-99a3-72eae3e41980-utilities" (OuterVolumeSpecName: "utilities") pod "285bc4f3-b69e-4c69-99a3-72eae3e41980" (UID: "285bc4f3-b69e-4c69-99a3-72eae3e41980"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.470434 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/285bc4f3-b69e-4c69-99a3-72eae3e41980-kube-api-access-q6vcx" (OuterVolumeSpecName: "kube-api-access-q6vcx") pod "285bc4f3-b69e-4c69-99a3-72eae3e41980" (UID: "285bc4f3-b69e-4c69-99a3-72eae3e41980"). InnerVolumeSpecName "kube-api-access-q6vcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.547209 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/285bc4f3-b69e-4c69-99a3-72eae3e41980-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "285bc4f3-b69e-4c69-99a3-72eae3e41980" (UID: "285bc4f3-b69e-4c69-99a3-72eae3e41980"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.559827 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/285bc4f3-b69e-4c69-99a3-72eae3e41980-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.559853 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6vcx\" (UniqueName: \"kubernetes.io/projected/285bc4f3-b69e-4c69-99a3-72eae3e41980-kube-api-access-q6vcx\") on node \"crc\" DevicePath \"\"" Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.559866 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/285bc4f3-b69e-4c69-99a3-72eae3e41980-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.833423 4745 generic.go:334] "Generic (PLEG): container finished" podID="285bc4f3-b69e-4c69-99a3-72eae3e41980" containerID="4388794e716a888de4a38d70eba5d80004b585de8d5c1457953240fd1e09ea33" exitCode=0 Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.833463 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcgxk" event={"ID":"285bc4f3-b69e-4c69-99a3-72eae3e41980","Type":"ContainerDied","Data":"4388794e716a888de4a38d70eba5d80004b585de8d5c1457953240fd1e09ea33"} Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.833488 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcgxk" event={"ID":"285bc4f3-b69e-4c69-99a3-72eae3e41980","Type":"ContainerDied","Data":"59327154f2bd4db7f41614698ab6d479aca568360ba840cb874aa13a1585300f"} Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.833507 4745 scope.go:117] "RemoveContainer" containerID="4388794e716a888de4a38d70eba5d80004b585de8d5c1457953240fd1e09ea33" Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.833552 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcgxk" Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.862651 4745 scope.go:117] "RemoveContainer" containerID="ab234fc267ff14c60024058124085b42b46306266a4b1006b4ee33d0c1c681e6" Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.867751 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qcgxk"] Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.876231 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qcgxk"] Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.883990 4745 scope.go:117] "RemoveContainer" containerID="c351e02f7c284aa20097380119bfda5f963cd368824be1fbd4e50f4efed6dc97" Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.946395 4745 scope.go:117] "RemoveContainer" containerID="4388794e716a888de4a38d70eba5d80004b585de8d5c1457953240fd1e09ea33" Oct 10 14:02:20 crc kubenswrapper[4745]: E1010 14:02:20.946799 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4388794e716a888de4a38d70eba5d80004b585de8d5c1457953240fd1e09ea33\": container with ID starting with 4388794e716a888de4a38d70eba5d80004b585de8d5c1457953240fd1e09ea33 not found: ID does not exist" containerID="4388794e716a888de4a38d70eba5d80004b585de8d5c1457953240fd1e09ea33" Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.946843 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4388794e716a888de4a38d70eba5d80004b585de8d5c1457953240fd1e09ea33"} err="failed to get container status \"4388794e716a888de4a38d70eba5d80004b585de8d5c1457953240fd1e09ea33\": rpc error: code = NotFound desc = could not find container \"4388794e716a888de4a38d70eba5d80004b585de8d5c1457953240fd1e09ea33\": container with ID starting with 4388794e716a888de4a38d70eba5d80004b585de8d5c1457953240fd1e09ea33 not found: ID does not exist" Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.946872 4745 scope.go:117] "RemoveContainer" containerID="ab234fc267ff14c60024058124085b42b46306266a4b1006b4ee33d0c1c681e6" Oct 10 14:02:20 crc kubenswrapper[4745]: E1010 14:02:20.947186 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab234fc267ff14c60024058124085b42b46306266a4b1006b4ee33d0c1c681e6\": container with ID starting with ab234fc267ff14c60024058124085b42b46306266a4b1006b4ee33d0c1c681e6 not found: ID does not exist" containerID="ab234fc267ff14c60024058124085b42b46306266a4b1006b4ee33d0c1c681e6" Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.947214 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab234fc267ff14c60024058124085b42b46306266a4b1006b4ee33d0c1c681e6"} err="failed to get container status \"ab234fc267ff14c60024058124085b42b46306266a4b1006b4ee33d0c1c681e6\": rpc error: code = NotFound desc = could not find container \"ab234fc267ff14c60024058124085b42b46306266a4b1006b4ee33d0c1c681e6\": container with ID starting with ab234fc267ff14c60024058124085b42b46306266a4b1006b4ee33d0c1c681e6 not found: ID does not exist" Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.947234 4745 scope.go:117] "RemoveContainer" containerID="c351e02f7c284aa20097380119bfda5f963cd368824be1fbd4e50f4efed6dc97" Oct 10 14:02:20 crc kubenswrapper[4745]: E1010 14:02:20.947394 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c351e02f7c284aa20097380119bfda5f963cd368824be1fbd4e50f4efed6dc97\": container with ID starting with c351e02f7c284aa20097380119bfda5f963cd368824be1fbd4e50f4efed6dc97 not found: ID does not exist" containerID="c351e02f7c284aa20097380119bfda5f963cd368824be1fbd4e50f4efed6dc97" Oct 10 14:02:20 crc kubenswrapper[4745]: I1010 14:02:20.947438 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c351e02f7c284aa20097380119bfda5f963cd368824be1fbd4e50f4efed6dc97"} err="failed to get container status \"c351e02f7c284aa20097380119bfda5f963cd368824be1fbd4e50f4efed6dc97\": rpc error: code = NotFound desc = could not find container \"c351e02f7c284aa20097380119bfda5f963cd368824be1fbd4e50f4efed6dc97\": container with ID starting with c351e02f7c284aa20097380119bfda5f963cd368824be1fbd4e50f4efed6dc97 not found: ID does not exist" Oct 10 14:02:22 crc kubenswrapper[4745]: I1010 14:02:22.768212 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="285bc4f3-b69e-4c69-99a3-72eae3e41980" path="/var/lib/kubelet/pods/285bc4f3-b69e-4c69-99a3-72eae3e41980/volumes" Oct 10 14:02:41 crc kubenswrapper[4745]: I1010 14:02:41.626632 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-csqs7"] Oct 10 14:02:41 crc kubenswrapper[4745]: E1010 14:02:41.627976 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="285bc4f3-b69e-4c69-99a3-72eae3e41980" containerName="registry-server" Oct 10 14:02:41 crc kubenswrapper[4745]: I1010 14:02:41.628002 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="285bc4f3-b69e-4c69-99a3-72eae3e41980" containerName="registry-server" Oct 10 14:02:41 crc kubenswrapper[4745]: E1010 14:02:41.628030 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="285bc4f3-b69e-4c69-99a3-72eae3e41980" containerName="extract-content" Oct 10 14:02:41 crc kubenswrapper[4745]: I1010 14:02:41.628043 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="285bc4f3-b69e-4c69-99a3-72eae3e41980" containerName="extract-content" Oct 10 14:02:41 crc kubenswrapper[4745]: E1010 14:02:41.628065 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="285bc4f3-b69e-4c69-99a3-72eae3e41980" containerName="extract-utilities" Oct 10 14:02:41 crc kubenswrapper[4745]: I1010 14:02:41.628080 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="285bc4f3-b69e-4c69-99a3-72eae3e41980" containerName="extract-utilities" Oct 10 14:02:41 crc kubenswrapper[4745]: I1010 14:02:41.628426 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="285bc4f3-b69e-4c69-99a3-72eae3e41980" containerName="registry-server" Oct 10 14:02:41 crc kubenswrapper[4745]: I1010 14:02:41.631136 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-csqs7" Oct 10 14:02:41 crc kubenswrapper[4745]: I1010 14:02:41.645331 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-csqs7"] Oct 10 14:02:41 crc kubenswrapper[4745]: I1010 14:02:41.826888 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9spjd\" (UniqueName: \"kubernetes.io/projected/e8e48b73-5d4d-4dad-91ff-99cabac012be-kube-api-access-9spjd\") pod \"redhat-marketplace-csqs7\" (UID: \"e8e48b73-5d4d-4dad-91ff-99cabac012be\") " pod="openshift-marketplace/redhat-marketplace-csqs7" Oct 10 14:02:41 crc kubenswrapper[4745]: I1010 14:02:41.827106 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8e48b73-5d4d-4dad-91ff-99cabac012be-catalog-content\") pod \"redhat-marketplace-csqs7\" (UID: \"e8e48b73-5d4d-4dad-91ff-99cabac012be\") " pod="openshift-marketplace/redhat-marketplace-csqs7" Oct 10 14:02:41 crc kubenswrapper[4745]: I1010 14:02:41.827482 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8e48b73-5d4d-4dad-91ff-99cabac012be-utilities\") pod \"redhat-marketplace-csqs7\" (UID: \"e8e48b73-5d4d-4dad-91ff-99cabac012be\") " pod="openshift-marketplace/redhat-marketplace-csqs7" Oct 10 14:02:41 crc kubenswrapper[4745]: I1010 14:02:41.929986 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9spjd\" (UniqueName: \"kubernetes.io/projected/e8e48b73-5d4d-4dad-91ff-99cabac012be-kube-api-access-9spjd\") pod \"redhat-marketplace-csqs7\" (UID: \"e8e48b73-5d4d-4dad-91ff-99cabac012be\") " pod="openshift-marketplace/redhat-marketplace-csqs7" Oct 10 14:02:41 crc kubenswrapper[4745]: I1010 14:02:41.930097 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8e48b73-5d4d-4dad-91ff-99cabac012be-catalog-content\") pod \"redhat-marketplace-csqs7\" (UID: \"e8e48b73-5d4d-4dad-91ff-99cabac012be\") " pod="openshift-marketplace/redhat-marketplace-csqs7" Oct 10 14:02:41 crc kubenswrapper[4745]: I1010 14:02:41.930202 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8e48b73-5d4d-4dad-91ff-99cabac012be-utilities\") pod \"redhat-marketplace-csqs7\" (UID: \"e8e48b73-5d4d-4dad-91ff-99cabac012be\") " pod="openshift-marketplace/redhat-marketplace-csqs7" Oct 10 14:02:41 crc kubenswrapper[4745]: I1010 14:02:41.930849 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8e48b73-5d4d-4dad-91ff-99cabac012be-utilities\") pod \"redhat-marketplace-csqs7\" (UID: \"e8e48b73-5d4d-4dad-91ff-99cabac012be\") " pod="openshift-marketplace/redhat-marketplace-csqs7" Oct 10 14:02:41 crc kubenswrapper[4745]: I1010 14:02:41.931180 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8e48b73-5d4d-4dad-91ff-99cabac012be-catalog-content\") pod \"redhat-marketplace-csqs7\" (UID: \"e8e48b73-5d4d-4dad-91ff-99cabac012be\") " pod="openshift-marketplace/redhat-marketplace-csqs7" Oct 10 14:02:41 crc kubenswrapper[4745]: I1010 14:02:41.953425 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9spjd\" (UniqueName: \"kubernetes.io/projected/e8e48b73-5d4d-4dad-91ff-99cabac012be-kube-api-access-9spjd\") pod \"redhat-marketplace-csqs7\" (UID: \"e8e48b73-5d4d-4dad-91ff-99cabac012be\") " pod="openshift-marketplace/redhat-marketplace-csqs7" Oct 10 14:02:41 crc kubenswrapper[4745]: I1010 14:02:41.965536 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-csqs7" Oct 10 14:02:42 crc kubenswrapper[4745]: I1010 14:02:42.443173 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-csqs7"] Oct 10 14:02:43 crc kubenswrapper[4745]: I1010 14:02:43.118436 4745 generic.go:334] "Generic (PLEG): container finished" podID="e8e48b73-5d4d-4dad-91ff-99cabac012be" containerID="b89dc22a079b0f7b49d8d543792712c2ebbc5d11d9b22f160db7d9ec96616d30" exitCode=0 Oct 10 14:02:43 crc kubenswrapper[4745]: I1010 14:02:43.118522 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-csqs7" event={"ID":"e8e48b73-5d4d-4dad-91ff-99cabac012be","Type":"ContainerDied","Data":"b89dc22a079b0f7b49d8d543792712c2ebbc5d11d9b22f160db7d9ec96616d30"} Oct 10 14:02:43 crc kubenswrapper[4745]: I1010 14:02:43.118805 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-csqs7" event={"ID":"e8e48b73-5d4d-4dad-91ff-99cabac012be","Type":"ContainerStarted","Data":"ec17877f139d39f014ddf1230ad838c2bd98a08527bbae34ace1c88a6c6f6633"} Oct 10 14:02:45 crc kubenswrapper[4745]: I1010 14:02:45.138757 4745 generic.go:334] "Generic (PLEG): container finished" podID="e8e48b73-5d4d-4dad-91ff-99cabac012be" containerID="02f914ede7fa5df7eeffba9a4eaadd3082dfd4be41ce9e3a27229656adcbefc4" exitCode=0 Oct 10 14:02:45 crc kubenswrapper[4745]: I1010 14:02:45.138831 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-csqs7" event={"ID":"e8e48b73-5d4d-4dad-91ff-99cabac012be","Type":"ContainerDied","Data":"02f914ede7fa5df7eeffba9a4eaadd3082dfd4be41ce9e3a27229656adcbefc4"} Oct 10 14:02:46 crc kubenswrapper[4745]: I1010 14:02:46.186853 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:02:46 crc kubenswrapper[4745]: I1010 14:02:46.187246 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:02:46 crc kubenswrapper[4745]: I1010 14:02:46.187300 4745 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 14:02:46 crc kubenswrapper[4745]: I1010 14:02:46.188166 4745 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1f503d509545908a6f3ef6fd56dd209acab0d1ff66704707b763337227936e6d"} pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 14:02:46 crc kubenswrapper[4745]: I1010 14:02:46.188245 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" containerID="cri-o://1f503d509545908a6f3ef6fd56dd209acab0d1ff66704707b763337227936e6d" gracePeriod=600 Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.161595 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-csqs7" event={"ID":"e8e48b73-5d4d-4dad-91ff-99cabac012be","Type":"ContainerStarted","Data":"b130728c16d6c94b872965ae7e9d58f22d7c3f50b2e97d19b1ba01ddb3129572"} Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.165975 4745 generic.go:334] "Generic (PLEG): container finished" podID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerID="1f503d509545908a6f3ef6fd56dd209acab0d1ff66704707b763337227936e6d" exitCode=0 Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.166025 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerDied","Data":"1f503d509545908a6f3ef6fd56dd209acab0d1ff66704707b763337227936e6d"} Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.166050 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c"} Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.166068 4745 scope.go:117] "RemoveContainer" containerID="631273c1f2726901c796455c92936db647302f2c051783bbcc0de5c3ba6d93c9" Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.192395 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-csqs7" podStartSLOduration=3.398011752 podStartE2EDuration="6.192358155s" podCreationTimestamp="2025-10-10 14:02:41 +0000 UTC" firstStartedPulling="2025-10-10 14:02:43.120615066 +0000 UTC m=+2677.018271869" lastFinishedPulling="2025-10-10 14:02:45.914961479 +0000 UTC m=+2679.812618272" observedRunningTime="2025-10-10 14:02:47.185364991 +0000 UTC m=+2681.083021754" watchObservedRunningTime="2025-10-10 14:02:47.192358155 +0000 UTC m=+2681.090014958" Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.819912 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.821608 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.823461 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-7q7m2" Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.825050 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.825247 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.826837 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.835179 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.956791 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.956876 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.956912 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/55dab8c7-31e2-4694-9f9e-a3272376cddd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.956942 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.956973 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55dab8c7-31e2-4694-9f9e-a3272376cddd-config-data\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.956989 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqpt4\" (UniqueName: \"kubernetes.io/projected/55dab8c7-31e2-4694-9f9e-a3272376cddd-kube-api-access-cqpt4\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.957011 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.957242 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/55dab8c7-31e2-4694-9f9e-a3272376cddd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:47 crc kubenswrapper[4745]: I1010 14:02:47.957331 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/55dab8c7-31e2-4694-9f9e-a3272376cddd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.058828 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.059252 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/55dab8c7-31e2-4694-9f9e-a3272376cddd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.059307 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/55dab8c7-31e2-4694-9f9e-a3272376cddd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.059484 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.059591 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.059632 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/55dab8c7-31e2-4694-9f9e-a3272376cddd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.059680 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.059747 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55dab8c7-31e2-4694-9f9e-a3272376cddd-config-data\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.059783 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqpt4\" (UniqueName: \"kubernetes.io/projected/55dab8c7-31e2-4694-9f9e-a3272376cddd-kube-api-access-cqpt4\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.059887 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.060025 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/55dab8c7-31e2-4694-9f9e-a3272376cddd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.060228 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/55dab8c7-31e2-4694-9f9e-a3272376cddd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.060619 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/55dab8c7-31e2-4694-9f9e-a3272376cddd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.061675 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55dab8c7-31e2-4694-9f9e-a3272376cddd-config-data\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.065749 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.071402 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.079484 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.082901 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqpt4\" (UniqueName: \"kubernetes.io/projected/55dab8c7-31e2-4694-9f9e-a3272376cddd-kube-api-access-cqpt4\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.098775 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.148374 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 10 14:02:48 crc kubenswrapper[4745]: I1010 14:02:48.626826 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 10 14:02:48 crc kubenswrapper[4745]: W1010 14:02:48.627845 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55dab8c7_31e2_4694_9f9e_a3272376cddd.slice/crio-22fe7f082cce9bd9583a7e8ee51ae1745f37fcb31e872d461c7497aa0c51e70e WatchSource:0}: Error finding container 22fe7f082cce9bd9583a7e8ee51ae1745f37fcb31e872d461c7497aa0c51e70e: Status 404 returned error can't find the container with id 22fe7f082cce9bd9583a7e8ee51ae1745f37fcb31e872d461c7497aa0c51e70e Oct 10 14:02:49 crc kubenswrapper[4745]: I1010 14:02:49.190598 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"55dab8c7-31e2-4694-9f9e-a3272376cddd","Type":"ContainerStarted","Data":"22fe7f082cce9bd9583a7e8ee51ae1745f37fcb31e872d461c7497aa0c51e70e"} Oct 10 14:02:51 crc kubenswrapper[4745]: I1010 14:02:51.965951 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-csqs7" Oct 10 14:02:51 crc kubenswrapper[4745]: I1010 14:02:51.966563 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-csqs7" Oct 10 14:02:52 crc kubenswrapper[4745]: I1010 14:02:52.018648 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-csqs7" Oct 10 14:02:52 crc kubenswrapper[4745]: I1010 14:02:52.268517 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-csqs7" Oct 10 14:02:52 crc kubenswrapper[4745]: I1010 14:02:52.307874 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-csqs7"] Oct 10 14:02:54 crc kubenswrapper[4745]: I1010 14:02:54.232422 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-csqs7" podUID="e8e48b73-5d4d-4dad-91ff-99cabac012be" containerName="registry-server" containerID="cri-o://b130728c16d6c94b872965ae7e9d58f22d7c3f50b2e97d19b1ba01ddb3129572" gracePeriod=2 Oct 10 14:02:54 crc kubenswrapper[4745]: I1010 14:02:54.682861 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-csqs7" Oct 10 14:02:54 crc kubenswrapper[4745]: I1010 14:02:54.788350 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8e48b73-5d4d-4dad-91ff-99cabac012be-catalog-content\") pod \"e8e48b73-5d4d-4dad-91ff-99cabac012be\" (UID: \"e8e48b73-5d4d-4dad-91ff-99cabac012be\") " Oct 10 14:02:54 crc kubenswrapper[4745]: I1010 14:02:54.788548 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9spjd\" (UniqueName: \"kubernetes.io/projected/e8e48b73-5d4d-4dad-91ff-99cabac012be-kube-api-access-9spjd\") pod \"e8e48b73-5d4d-4dad-91ff-99cabac012be\" (UID: \"e8e48b73-5d4d-4dad-91ff-99cabac012be\") " Oct 10 14:02:54 crc kubenswrapper[4745]: I1010 14:02:54.788624 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8e48b73-5d4d-4dad-91ff-99cabac012be-utilities\") pod \"e8e48b73-5d4d-4dad-91ff-99cabac012be\" (UID: \"e8e48b73-5d4d-4dad-91ff-99cabac012be\") " Oct 10 14:02:54 crc kubenswrapper[4745]: I1010 14:02:54.790097 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8e48b73-5d4d-4dad-91ff-99cabac012be-utilities" (OuterVolumeSpecName: "utilities") pod "e8e48b73-5d4d-4dad-91ff-99cabac012be" (UID: "e8e48b73-5d4d-4dad-91ff-99cabac012be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:02:54 crc kubenswrapper[4745]: I1010 14:02:54.804375 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8e48b73-5d4d-4dad-91ff-99cabac012be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8e48b73-5d4d-4dad-91ff-99cabac012be" (UID: "e8e48b73-5d4d-4dad-91ff-99cabac012be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:02:54 crc kubenswrapper[4745]: I1010 14:02:54.806434 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8e48b73-5d4d-4dad-91ff-99cabac012be-kube-api-access-9spjd" (OuterVolumeSpecName: "kube-api-access-9spjd") pod "e8e48b73-5d4d-4dad-91ff-99cabac012be" (UID: "e8e48b73-5d4d-4dad-91ff-99cabac012be"). InnerVolumeSpecName "kube-api-access-9spjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:02:54 crc kubenswrapper[4745]: I1010 14:02:54.891556 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8e48b73-5d4d-4dad-91ff-99cabac012be-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:02:54 crc kubenswrapper[4745]: I1010 14:02:54.891598 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9spjd\" (UniqueName: \"kubernetes.io/projected/e8e48b73-5d4d-4dad-91ff-99cabac012be-kube-api-access-9spjd\") on node \"crc\" DevicePath \"\"" Oct 10 14:02:54 crc kubenswrapper[4745]: I1010 14:02:54.891614 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8e48b73-5d4d-4dad-91ff-99cabac012be-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:02:55 crc kubenswrapper[4745]: I1010 14:02:55.241623 4745 generic.go:334] "Generic (PLEG): container finished" podID="e8e48b73-5d4d-4dad-91ff-99cabac012be" containerID="b130728c16d6c94b872965ae7e9d58f22d7c3f50b2e97d19b1ba01ddb3129572" exitCode=0 Oct 10 14:02:55 crc kubenswrapper[4745]: I1010 14:02:55.241886 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-csqs7" event={"ID":"e8e48b73-5d4d-4dad-91ff-99cabac012be","Type":"ContainerDied","Data":"b130728c16d6c94b872965ae7e9d58f22d7c3f50b2e97d19b1ba01ddb3129572"} Oct 10 14:02:55 crc kubenswrapper[4745]: I1010 14:02:55.241911 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-csqs7" event={"ID":"e8e48b73-5d4d-4dad-91ff-99cabac012be","Type":"ContainerDied","Data":"ec17877f139d39f014ddf1230ad838c2bd98a08527bbae34ace1c88a6c6f6633"} Oct 10 14:02:55 crc kubenswrapper[4745]: I1010 14:02:55.241927 4745 scope.go:117] "RemoveContainer" containerID="b130728c16d6c94b872965ae7e9d58f22d7c3f50b2e97d19b1ba01ddb3129572" Oct 10 14:02:55 crc kubenswrapper[4745]: I1010 14:02:55.242039 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-csqs7" Oct 10 14:02:55 crc kubenswrapper[4745]: I1010 14:02:55.278122 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-csqs7"] Oct 10 14:02:55 crc kubenswrapper[4745]: I1010 14:02:55.285767 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-csqs7"] Oct 10 14:02:55 crc kubenswrapper[4745]: I1010 14:02:55.287345 4745 scope.go:117] "RemoveContainer" containerID="02f914ede7fa5df7eeffba9a4eaadd3082dfd4be41ce9e3a27229656adcbefc4" Oct 10 14:02:55 crc kubenswrapper[4745]: I1010 14:02:55.305965 4745 scope.go:117] "RemoveContainer" containerID="b89dc22a079b0f7b49d8d543792712c2ebbc5d11d9b22f160db7d9ec96616d30" Oct 10 14:02:55 crc kubenswrapper[4745]: I1010 14:02:55.347142 4745 scope.go:117] "RemoveContainer" containerID="b130728c16d6c94b872965ae7e9d58f22d7c3f50b2e97d19b1ba01ddb3129572" Oct 10 14:02:55 crc kubenswrapper[4745]: E1010 14:02:55.347608 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b130728c16d6c94b872965ae7e9d58f22d7c3f50b2e97d19b1ba01ddb3129572\": container with ID starting with b130728c16d6c94b872965ae7e9d58f22d7c3f50b2e97d19b1ba01ddb3129572 not found: ID does not exist" containerID="b130728c16d6c94b872965ae7e9d58f22d7c3f50b2e97d19b1ba01ddb3129572" Oct 10 14:02:55 crc kubenswrapper[4745]: I1010 14:02:55.347649 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b130728c16d6c94b872965ae7e9d58f22d7c3f50b2e97d19b1ba01ddb3129572"} err="failed to get container status \"b130728c16d6c94b872965ae7e9d58f22d7c3f50b2e97d19b1ba01ddb3129572\": rpc error: code = NotFound desc = could not find container \"b130728c16d6c94b872965ae7e9d58f22d7c3f50b2e97d19b1ba01ddb3129572\": container with ID starting with b130728c16d6c94b872965ae7e9d58f22d7c3f50b2e97d19b1ba01ddb3129572 not found: ID does not exist" Oct 10 14:02:55 crc kubenswrapper[4745]: I1010 14:02:55.347683 4745 scope.go:117] "RemoveContainer" containerID="02f914ede7fa5df7eeffba9a4eaadd3082dfd4be41ce9e3a27229656adcbefc4" Oct 10 14:02:55 crc kubenswrapper[4745]: E1010 14:02:55.347984 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02f914ede7fa5df7eeffba9a4eaadd3082dfd4be41ce9e3a27229656adcbefc4\": container with ID starting with 02f914ede7fa5df7eeffba9a4eaadd3082dfd4be41ce9e3a27229656adcbefc4 not found: ID does not exist" containerID="02f914ede7fa5df7eeffba9a4eaadd3082dfd4be41ce9e3a27229656adcbefc4" Oct 10 14:02:55 crc kubenswrapper[4745]: I1010 14:02:55.348035 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02f914ede7fa5df7eeffba9a4eaadd3082dfd4be41ce9e3a27229656adcbefc4"} err="failed to get container status \"02f914ede7fa5df7eeffba9a4eaadd3082dfd4be41ce9e3a27229656adcbefc4\": rpc error: code = NotFound desc = could not find container \"02f914ede7fa5df7eeffba9a4eaadd3082dfd4be41ce9e3a27229656adcbefc4\": container with ID starting with 02f914ede7fa5df7eeffba9a4eaadd3082dfd4be41ce9e3a27229656adcbefc4 not found: ID does not exist" Oct 10 14:02:55 crc kubenswrapper[4745]: I1010 14:02:55.348065 4745 scope.go:117] "RemoveContainer" containerID="b89dc22a079b0f7b49d8d543792712c2ebbc5d11d9b22f160db7d9ec96616d30" Oct 10 14:02:55 crc kubenswrapper[4745]: E1010 14:02:55.348451 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b89dc22a079b0f7b49d8d543792712c2ebbc5d11d9b22f160db7d9ec96616d30\": container with ID starting with b89dc22a079b0f7b49d8d543792712c2ebbc5d11d9b22f160db7d9ec96616d30 not found: ID does not exist" containerID="b89dc22a079b0f7b49d8d543792712c2ebbc5d11d9b22f160db7d9ec96616d30" Oct 10 14:02:55 crc kubenswrapper[4745]: I1010 14:02:55.348477 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b89dc22a079b0f7b49d8d543792712c2ebbc5d11d9b22f160db7d9ec96616d30"} err="failed to get container status \"b89dc22a079b0f7b49d8d543792712c2ebbc5d11d9b22f160db7d9ec96616d30\": rpc error: code = NotFound desc = could not find container \"b89dc22a079b0f7b49d8d543792712c2ebbc5d11d9b22f160db7d9ec96616d30\": container with ID starting with b89dc22a079b0f7b49d8d543792712c2ebbc5d11d9b22f160db7d9ec96616d30 not found: ID does not exist" Oct 10 14:02:56 crc kubenswrapper[4745]: I1010 14:02:56.759169 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8e48b73-5d4d-4dad-91ff-99cabac012be" path="/var/lib/kubelet/pods/e8e48b73-5d4d-4dad-91ff-99cabac012be/volumes" Oct 10 14:03:04 crc kubenswrapper[4745]: I1010 14:03:04.834789 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hh4lq"] Oct 10 14:03:04 crc kubenswrapper[4745]: E1010 14:03:04.835891 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8e48b73-5d4d-4dad-91ff-99cabac012be" containerName="registry-server" Oct 10 14:03:04 crc kubenswrapper[4745]: I1010 14:03:04.835904 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8e48b73-5d4d-4dad-91ff-99cabac012be" containerName="registry-server" Oct 10 14:03:04 crc kubenswrapper[4745]: E1010 14:03:04.835914 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8e48b73-5d4d-4dad-91ff-99cabac012be" containerName="extract-utilities" Oct 10 14:03:04 crc kubenswrapper[4745]: I1010 14:03:04.835921 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8e48b73-5d4d-4dad-91ff-99cabac012be" containerName="extract-utilities" Oct 10 14:03:04 crc kubenswrapper[4745]: E1010 14:03:04.835932 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8e48b73-5d4d-4dad-91ff-99cabac012be" containerName="extract-content" Oct 10 14:03:04 crc kubenswrapper[4745]: I1010 14:03:04.835937 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8e48b73-5d4d-4dad-91ff-99cabac012be" containerName="extract-content" Oct 10 14:03:04 crc kubenswrapper[4745]: I1010 14:03:04.836133 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8e48b73-5d4d-4dad-91ff-99cabac012be" containerName="registry-server" Oct 10 14:03:04 crc kubenswrapper[4745]: I1010 14:03:04.837483 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hh4lq" Oct 10 14:03:04 crc kubenswrapper[4745]: I1010 14:03:04.852253 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hh4lq"] Oct 10 14:03:04 crc kubenswrapper[4745]: I1010 14:03:04.943786 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d974131-07e9-4746-aca2-69e3325c062b-utilities\") pod \"certified-operators-hh4lq\" (UID: \"3d974131-07e9-4746-aca2-69e3325c062b\") " pod="openshift-marketplace/certified-operators-hh4lq" Oct 10 14:03:04 crc kubenswrapper[4745]: I1010 14:03:04.943942 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k5qn\" (UniqueName: \"kubernetes.io/projected/3d974131-07e9-4746-aca2-69e3325c062b-kube-api-access-2k5qn\") pod \"certified-operators-hh4lq\" (UID: \"3d974131-07e9-4746-aca2-69e3325c062b\") " pod="openshift-marketplace/certified-operators-hh4lq" Oct 10 14:03:04 crc kubenswrapper[4745]: I1010 14:03:04.944118 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d974131-07e9-4746-aca2-69e3325c062b-catalog-content\") pod \"certified-operators-hh4lq\" (UID: \"3d974131-07e9-4746-aca2-69e3325c062b\") " pod="openshift-marketplace/certified-operators-hh4lq" Oct 10 14:03:05 crc kubenswrapper[4745]: I1010 14:03:05.045892 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d974131-07e9-4746-aca2-69e3325c062b-catalog-content\") pod \"certified-operators-hh4lq\" (UID: \"3d974131-07e9-4746-aca2-69e3325c062b\") " pod="openshift-marketplace/certified-operators-hh4lq" Oct 10 14:03:05 crc kubenswrapper[4745]: I1010 14:03:05.046241 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d974131-07e9-4746-aca2-69e3325c062b-utilities\") pod \"certified-operators-hh4lq\" (UID: \"3d974131-07e9-4746-aca2-69e3325c062b\") " pod="openshift-marketplace/certified-operators-hh4lq" Oct 10 14:03:05 crc kubenswrapper[4745]: I1010 14:03:05.046289 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k5qn\" (UniqueName: \"kubernetes.io/projected/3d974131-07e9-4746-aca2-69e3325c062b-kube-api-access-2k5qn\") pod \"certified-operators-hh4lq\" (UID: \"3d974131-07e9-4746-aca2-69e3325c062b\") " pod="openshift-marketplace/certified-operators-hh4lq" Oct 10 14:03:05 crc kubenswrapper[4745]: I1010 14:03:05.046397 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d974131-07e9-4746-aca2-69e3325c062b-catalog-content\") pod \"certified-operators-hh4lq\" (UID: \"3d974131-07e9-4746-aca2-69e3325c062b\") " pod="openshift-marketplace/certified-operators-hh4lq" Oct 10 14:03:05 crc kubenswrapper[4745]: I1010 14:03:05.047413 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d974131-07e9-4746-aca2-69e3325c062b-utilities\") pod \"certified-operators-hh4lq\" (UID: \"3d974131-07e9-4746-aca2-69e3325c062b\") " pod="openshift-marketplace/certified-operators-hh4lq" Oct 10 14:03:05 crc kubenswrapper[4745]: I1010 14:03:05.074827 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k5qn\" (UniqueName: \"kubernetes.io/projected/3d974131-07e9-4746-aca2-69e3325c062b-kube-api-access-2k5qn\") pod \"certified-operators-hh4lq\" (UID: \"3d974131-07e9-4746-aca2-69e3325c062b\") " pod="openshift-marketplace/certified-operators-hh4lq" Oct 10 14:03:05 crc kubenswrapper[4745]: I1010 14:03:05.177716 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hh4lq" Oct 10 14:03:22 crc kubenswrapper[4745]: E1010 14:03:22.620964 4745 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 10 14:03:22 crc kubenswrapper[4745]: E1010 14:03:22.621862 4745 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cqpt4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(55dab8c7-31e2-4694-9f9e-a3272376cddd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 14:03:22 crc kubenswrapper[4745]: E1010 14:03:22.623099 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="55dab8c7-31e2-4694-9f9e-a3272376cddd" Oct 10 14:03:23 crc kubenswrapper[4745]: I1010 14:03:23.065204 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hh4lq"] Oct 10 14:03:23 crc kubenswrapper[4745]: I1010 14:03:23.546024 4745 generic.go:334] "Generic (PLEG): container finished" podID="3d974131-07e9-4746-aca2-69e3325c062b" containerID="0f70d5edeeae0697f4bd160685f97552ab412f135deda4b5510414a55e8445cf" exitCode=0 Oct 10 14:03:23 crc kubenswrapper[4745]: I1010 14:03:23.546107 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh4lq" event={"ID":"3d974131-07e9-4746-aca2-69e3325c062b","Type":"ContainerDied","Data":"0f70d5edeeae0697f4bd160685f97552ab412f135deda4b5510414a55e8445cf"} Oct 10 14:03:23 crc kubenswrapper[4745]: I1010 14:03:23.546406 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh4lq" event={"ID":"3d974131-07e9-4746-aca2-69e3325c062b","Type":"ContainerStarted","Data":"2ef8c9c7f926783965807a095359574e816a97ebd9c812365e6adaca3cf66853"} Oct 10 14:03:23 crc kubenswrapper[4745]: E1010 14:03:23.548012 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="55dab8c7-31e2-4694-9f9e-a3272376cddd" Oct 10 14:03:25 crc kubenswrapper[4745]: I1010 14:03:25.584113 4745 generic.go:334] "Generic (PLEG): container finished" podID="3d974131-07e9-4746-aca2-69e3325c062b" containerID="f60a566889df4814ec93ea79d50e26595e1142e46cf3e67f1f67592a180fa9d0" exitCode=0 Oct 10 14:03:25 crc kubenswrapper[4745]: I1010 14:03:25.584194 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh4lq" event={"ID":"3d974131-07e9-4746-aca2-69e3325c062b","Type":"ContainerDied","Data":"f60a566889df4814ec93ea79d50e26595e1142e46cf3e67f1f67592a180fa9d0"} Oct 10 14:03:26 crc kubenswrapper[4745]: I1010 14:03:26.598768 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh4lq" event={"ID":"3d974131-07e9-4746-aca2-69e3325c062b","Type":"ContainerStarted","Data":"601b0c1127d1ce178632c3a7c1645b06950366b51fda54a8ca384e490b3d165b"} Oct 10 14:03:26 crc kubenswrapper[4745]: I1010 14:03:26.627245 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hh4lq" podStartSLOduration=20.107349049 podStartE2EDuration="22.627226621s" podCreationTimestamp="2025-10-10 14:03:04 +0000 UTC" firstStartedPulling="2025-10-10 14:03:23.548320686 +0000 UTC m=+2717.445977449" lastFinishedPulling="2025-10-10 14:03:26.068198248 +0000 UTC m=+2719.965855021" observedRunningTime="2025-10-10 14:03:26.624639157 +0000 UTC m=+2720.522295960" watchObservedRunningTime="2025-10-10 14:03:26.627226621 +0000 UTC m=+2720.524883394" Oct 10 14:03:35 crc kubenswrapper[4745]: I1010 14:03:35.178122 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hh4lq" Oct 10 14:03:35 crc kubenswrapper[4745]: I1010 14:03:35.178948 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hh4lq" Oct 10 14:03:35 crc kubenswrapper[4745]: I1010 14:03:35.245833 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hh4lq" Oct 10 14:03:35 crc kubenswrapper[4745]: I1010 14:03:35.744161 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hh4lq" Oct 10 14:03:36 crc kubenswrapper[4745]: I1010 14:03:36.030920 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hh4lq"] Oct 10 14:03:37 crc kubenswrapper[4745]: I1010 14:03:37.714391 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hh4lq" podUID="3d974131-07e9-4746-aca2-69e3325c062b" containerName="registry-server" containerID="cri-o://601b0c1127d1ce178632c3a7c1645b06950366b51fda54a8ca384e490b3d165b" gracePeriod=2 Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.250580 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hh4lq" Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.340972 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d974131-07e9-4746-aca2-69e3325c062b-catalog-content\") pod \"3d974131-07e9-4746-aca2-69e3325c062b\" (UID: \"3d974131-07e9-4746-aca2-69e3325c062b\") " Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.388954 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d974131-07e9-4746-aca2-69e3325c062b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d974131-07e9-4746-aca2-69e3325c062b" (UID: "3d974131-07e9-4746-aca2-69e3325c062b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.442429 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k5qn\" (UniqueName: \"kubernetes.io/projected/3d974131-07e9-4746-aca2-69e3325c062b-kube-api-access-2k5qn\") pod \"3d974131-07e9-4746-aca2-69e3325c062b\" (UID: \"3d974131-07e9-4746-aca2-69e3325c062b\") " Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.442510 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d974131-07e9-4746-aca2-69e3325c062b-utilities\") pod \"3d974131-07e9-4746-aca2-69e3325c062b\" (UID: \"3d974131-07e9-4746-aca2-69e3325c062b\") " Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.443056 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d974131-07e9-4746-aca2-69e3325c062b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.448290 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d974131-07e9-4746-aca2-69e3325c062b-utilities" (OuterVolumeSpecName: "utilities") pod "3d974131-07e9-4746-aca2-69e3325c062b" (UID: "3d974131-07e9-4746-aca2-69e3325c062b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.461853 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d974131-07e9-4746-aca2-69e3325c062b-kube-api-access-2k5qn" (OuterVolumeSpecName: "kube-api-access-2k5qn") pod "3d974131-07e9-4746-aca2-69e3325c062b" (UID: "3d974131-07e9-4746-aca2-69e3325c062b"). InnerVolumeSpecName "kube-api-access-2k5qn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.545951 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k5qn\" (UniqueName: \"kubernetes.io/projected/3d974131-07e9-4746-aca2-69e3325c062b-kube-api-access-2k5qn\") on node \"crc\" DevicePath \"\"" Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.546017 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d974131-07e9-4746-aca2-69e3325c062b-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.729167 4745 generic.go:334] "Generic (PLEG): container finished" podID="3d974131-07e9-4746-aca2-69e3325c062b" containerID="601b0c1127d1ce178632c3a7c1645b06950366b51fda54a8ca384e490b3d165b" exitCode=0 Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.729255 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh4lq" event={"ID":"3d974131-07e9-4746-aca2-69e3325c062b","Type":"ContainerDied","Data":"601b0c1127d1ce178632c3a7c1645b06950366b51fda54a8ca384e490b3d165b"} Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.729279 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hh4lq" Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.729310 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh4lq" event={"ID":"3d974131-07e9-4746-aca2-69e3325c062b","Type":"ContainerDied","Data":"2ef8c9c7f926783965807a095359574e816a97ebd9c812365e6adaca3cf66853"} Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.729342 4745 scope.go:117] "RemoveContainer" containerID="601b0c1127d1ce178632c3a7c1645b06950366b51fda54a8ca384e490b3d165b" Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.761819 4745 scope.go:117] "RemoveContainer" containerID="f60a566889df4814ec93ea79d50e26595e1142e46cf3e67f1f67592a180fa9d0" Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.802129 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hh4lq"] Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.802467 4745 scope.go:117] "RemoveContainer" containerID="0f70d5edeeae0697f4bd160685f97552ab412f135deda4b5510414a55e8445cf" Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.812774 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hh4lq"] Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.844917 4745 scope.go:117] "RemoveContainer" containerID="601b0c1127d1ce178632c3a7c1645b06950366b51fda54a8ca384e490b3d165b" Oct 10 14:03:38 crc kubenswrapper[4745]: E1010 14:03:38.845395 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"601b0c1127d1ce178632c3a7c1645b06950366b51fda54a8ca384e490b3d165b\": container with ID starting with 601b0c1127d1ce178632c3a7c1645b06950366b51fda54a8ca384e490b3d165b not found: ID does not exist" containerID="601b0c1127d1ce178632c3a7c1645b06950366b51fda54a8ca384e490b3d165b" Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.845425 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"601b0c1127d1ce178632c3a7c1645b06950366b51fda54a8ca384e490b3d165b"} err="failed to get container status \"601b0c1127d1ce178632c3a7c1645b06950366b51fda54a8ca384e490b3d165b\": rpc error: code = NotFound desc = could not find container \"601b0c1127d1ce178632c3a7c1645b06950366b51fda54a8ca384e490b3d165b\": container with ID starting with 601b0c1127d1ce178632c3a7c1645b06950366b51fda54a8ca384e490b3d165b not found: ID does not exist" Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.845450 4745 scope.go:117] "RemoveContainer" containerID="f60a566889df4814ec93ea79d50e26595e1142e46cf3e67f1f67592a180fa9d0" Oct 10 14:03:38 crc kubenswrapper[4745]: E1010 14:03:38.845741 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f60a566889df4814ec93ea79d50e26595e1142e46cf3e67f1f67592a180fa9d0\": container with ID starting with f60a566889df4814ec93ea79d50e26595e1142e46cf3e67f1f67592a180fa9d0 not found: ID does not exist" containerID="f60a566889df4814ec93ea79d50e26595e1142e46cf3e67f1f67592a180fa9d0" Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.845761 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f60a566889df4814ec93ea79d50e26595e1142e46cf3e67f1f67592a180fa9d0"} err="failed to get container status \"f60a566889df4814ec93ea79d50e26595e1142e46cf3e67f1f67592a180fa9d0\": rpc error: code = NotFound desc = could not find container \"f60a566889df4814ec93ea79d50e26595e1142e46cf3e67f1f67592a180fa9d0\": container with ID starting with f60a566889df4814ec93ea79d50e26595e1142e46cf3e67f1f67592a180fa9d0 not found: ID does not exist" Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.845775 4745 scope.go:117] "RemoveContainer" containerID="0f70d5edeeae0697f4bd160685f97552ab412f135deda4b5510414a55e8445cf" Oct 10 14:03:38 crc kubenswrapper[4745]: E1010 14:03:38.846302 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f70d5edeeae0697f4bd160685f97552ab412f135deda4b5510414a55e8445cf\": container with ID starting with 0f70d5edeeae0697f4bd160685f97552ab412f135deda4b5510414a55e8445cf not found: ID does not exist" containerID="0f70d5edeeae0697f4bd160685f97552ab412f135deda4b5510414a55e8445cf" Oct 10 14:03:38 crc kubenswrapper[4745]: I1010 14:03:38.846369 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f70d5edeeae0697f4bd160685f97552ab412f135deda4b5510414a55e8445cf"} err="failed to get container status \"0f70d5edeeae0697f4bd160685f97552ab412f135deda4b5510414a55e8445cf\": rpc error: code = NotFound desc = could not find container \"0f70d5edeeae0697f4bd160685f97552ab412f135deda4b5510414a55e8445cf\": container with ID starting with 0f70d5edeeae0697f4bd160685f97552ab412f135deda4b5510414a55e8445cf not found: ID does not exist" Oct 10 14:03:39 crc kubenswrapper[4745]: I1010 14:03:39.523971 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 10 14:03:40 crc kubenswrapper[4745]: I1010 14:03:40.757365 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d974131-07e9-4746-aca2-69e3325c062b" path="/var/lib/kubelet/pods/3d974131-07e9-4746-aca2-69e3325c062b/volumes" Oct 10 14:03:40 crc kubenswrapper[4745]: I1010 14:03:40.760242 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"55dab8c7-31e2-4694-9f9e-a3272376cddd","Type":"ContainerStarted","Data":"3614647b235f87652bbfd4290e8f51f2c04d41c401b7a1827317a48b499da92f"} Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.513496 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=48.622168448 podStartE2EDuration="1m39.513460228s" podCreationTimestamp="2025-10-10 14:02:46 +0000 UTC" firstStartedPulling="2025-10-10 14:02:48.630089975 +0000 UTC m=+2682.527746728" lastFinishedPulling="2025-10-10 14:03:39.521381745 +0000 UTC m=+2733.419038508" observedRunningTime="2025-10-10 14:03:40.784100036 +0000 UTC m=+2734.681756809" watchObservedRunningTime="2025-10-10 14:04:25.513460228 +0000 UTC m=+2779.411117021" Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.519636 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9gtbx"] Oct 10 14:04:25 crc kubenswrapper[4745]: E1010 14:04:25.520350 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d974131-07e9-4746-aca2-69e3325c062b" containerName="extract-utilities" Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.520385 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d974131-07e9-4746-aca2-69e3325c062b" containerName="extract-utilities" Oct 10 14:04:25 crc kubenswrapper[4745]: E1010 14:04:25.520442 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d974131-07e9-4746-aca2-69e3325c062b" containerName="registry-server" Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.520457 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d974131-07e9-4746-aca2-69e3325c062b" containerName="registry-server" Oct 10 14:04:25 crc kubenswrapper[4745]: E1010 14:04:25.520490 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d974131-07e9-4746-aca2-69e3325c062b" containerName="extract-content" Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.520504 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d974131-07e9-4746-aca2-69e3325c062b" containerName="extract-content" Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.520907 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d974131-07e9-4746-aca2-69e3325c062b" containerName="registry-server" Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.523508 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9gtbx" Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.535984 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9gtbx"] Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.688584 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqncg\" (UniqueName: \"kubernetes.io/projected/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-kube-api-access-pqncg\") pod \"community-operators-9gtbx\" (UID: \"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e\") " pod="openshift-marketplace/community-operators-9gtbx" Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.688692 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-catalog-content\") pod \"community-operators-9gtbx\" (UID: \"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e\") " pod="openshift-marketplace/community-operators-9gtbx" Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.688877 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-utilities\") pod \"community-operators-9gtbx\" (UID: \"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e\") " pod="openshift-marketplace/community-operators-9gtbx" Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.790261 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-utilities\") pod \"community-operators-9gtbx\" (UID: \"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e\") " pod="openshift-marketplace/community-operators-9gtbx" Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.790364 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqncg\" (UniqueName: \"kubernetes.io/projected/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-kube-api-access-pqncg\") pod \"community-operators-9gtbx\" (UID: \"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e\") " pod="openshift-marketplace/community-operators-9gtbx" Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.790406 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-catalog-content\") pod \"community-operators-9gtbx\" (UID: \"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e\") " pod="openshift-marketplace/community-operators-9gtbx" Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.790861 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-catalog-content\") pod \"community-operators-9gtbx\" (UID: \"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e\") " pod="openshift-marketplace/community-operators-9gtbx" Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.791436 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-utilities\") pod \"community-operators-9gtbx\" (UID: \"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e\") " pod="openshift-marketplace/community-operators-9gtbx" Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.811997 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqncg\" (UniqueName: \"kubernetes.io/projected/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-kube-api-access-pqncg\") pod \"community-operators-9gtbx\" (UID: \"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e\") " pod="openshift-marketplace/community-operators-9gtbx" Oct 10 14:04:25 crc kubenswrapper[4745]: I1010 14:04:25.888505 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9gtbx" Oct 10 14:04:26 crc kubenswrapper[4745]: I1010 14:04:26.360650 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9gtbx"] Oct 10 14:04:27 crc kubenswrapper[4745]: I1010 14:04:27.267783 4745 generic.go:334] "Generic (PLEG): container finished" podID="b5e45037-8834-458d-b1e9-fa2b8fc7dc8e" containerID="e0a2bb8866b2a49bf35b23f2cf016df81bd1a86938a360ca03f8d591bfc6bbcd" exitCode=0 Oct 10 14:04:27 crc kubenswrapper[4745]: I1010 14:04:27.268246 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gtbx" event={"ID":"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e","Type":"ContainerDied","Data":"e0a2bb8866b2a49bf35b23f2cf016df81bd1a86938a360ca03f8d591bfc6bbcd"} Oct 10 14:04:27 crc kubenswrapper[4745]: I1010 14:04:27.268904 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gtbx" event={"ID":"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e","Type":"ContainerStarted","Data":"f18874e6df55356bcb5e1583bd1408b2e4312f0f94eca9d54e04f8402497a275"} Oct 10 14:04:29 crc kubenswrapper[4745]: I1010 14:04:29.293713 4745 generic.go:334] "Generic (PLEG): container finished" podID="b5e45037-8834-458d-b1e9-fa2b8fc7dc8e" containerID="b08c891927d1d9b43721ab5f99d80afc8bcdf78bcb7c87f13e350b82237d2fab" exitCode=0 Oct 10 14:04:29 crc kubenswrapper[4745]: I1010 14:04:29.293821 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gtbx" event={"ID":"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e","Type":"ContainerDied","Data":"b08c891927d1d9b43721ab5f99d80afc8bcdf78bcb7c87f13e350b82237d2fab"} Oct 10 14:04:30 crc kubenswrapper[4745]: I1010 14:04:30.319554 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gtbx" event={"ID":"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e","Type":"ContainerStarted","Data":"86ea154fc72c12ce2865c12f27e9cac327954dead185f22d2afda93851693cd2"} Oct 10 14:04:30 crc kubenswrapper[4745]: I1010 14:04:30.341631 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9gtbx" podStartSLOduration=2.684223584 podStartE2EDuration="5.341608413s" podCreationTimestamp="2025-10-10 14:04:25 +0000 UTC" firstStartedPulling="2025-10-10 14:04:27.271120038 +0000 UTC m=+2781.168776801" lastFinishedPulling="2025-10-10 14:04:29.928504867 +0000 UTC m=+2783.826161630" observedRunningTime="2025-10-10 14:04:30.339913541 +0000 UTC m=+2784.237570314" watchObservedRunningTime="2025-10-10 14:04:30.341608413 +0000 UTC m=+2784.239265196" Oct 10 14:04:35 crc kubenswrapper[4745]: I1010 14:04:35.889470 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9gtbx" Oct 10 14:04:35 crc kubenswrapper[4745]: I1010 14:04:35.890166 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9gtbx" Oct 10 14:04:35 crc kubenswrapper[4745]: I1010 14:04:35.950153 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9gtbx" Oct 10 14:04:36 crc kubenswrapper[4745]: I1010 14:04:36.416609 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9gtbx" Oct 10 14:04:36 crc kubenswrapper[4745]: I1010 14:04:36.471450 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9gtbx"] Oct 10 14:04:38 crc kubenswrapper[4745]: I1010 14:04:38.389596 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9gtbx" podUID="b5e45037-8834-458d-b1e9-fa2b8fc7dc8e" containerName="registry-server" containerID="cri-o://86ea154fc72c12ce2865c12f27e9cac327954dead185f22d2afda93851693cd2" gracePeriod=2 Oct 10 14:04:38 crc kubenswrapper[4745]: I1010 14:04:38.967943 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9gtbx" Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.072001 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-catalog-content\") pod \"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e\" (UID: \"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e\") " Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.072123 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqncg\" (UniqueName: \"kubernetes.io/projected/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-kube-api-access-pqncg\") pod \"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e\" (UID: \"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e\") " Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.072187 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-utilities\") pod \"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e\" (UID: \"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e\") " Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.073171 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-utilities" (OuterVolumeSpecName: "utilities") pod "b5e45037-8834-458d-b1e9-fa2b8fc7dc8e" (UID: "b5e45037-8834-458d-b1e9-fa2b8fc7dc8e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.081079 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-kube-api-access-pqncg" (OuterVolumeSpecName: "kube-api-access-pqncg") pod "b5e45037-8834-458d-b1e9-fa2b8fc7dc8e" (UID: "b5e45037-8834-458d-b1e9-fa2b8fc7dc8e"). InnerVolumeSpecName "kube-api-access-pqncg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.134248 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5e45037-8834-458d-b1e9-fa2b8fc7dc8e" (UID: "b5e45037-8834-458d-b1e9-fa2b8fc7dc8e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.174935 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqncg\" (UniqueName: \"kubernetes.io/projected/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-kube-api-access-pqncg\") on node \"crc\" DevicePath \"\"" Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.174966 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.174979 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.401298 4745 generic.go:334] "Generic (PLEG): container finished" podID="b5e45037-8834-458d-b1e9-fa2b8fc7dc8e" containerID="86ea154fc72c12ce2865c12f27e9cac327954dead185f22d2afda93851693cd2" exitCode=0 Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.401350 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gtbx" event={"ID":"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e","Type":"ContainerDied","Data":"86ea154fc72c12ce2865c12f27e9cac327954dead185f22d2afda93851693cd2"} Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.401371 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9gtbx" Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.401392 4745 scope.go:117] "RemoveContainer" containerID="86ea154fc72c12ce2865c12f27e9cac327954dead185f22d2afda93851693cd2" Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.401379 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gtbx" event={"ID":"b5e45037-8834-458d-b1e9-fa2b8fc7dc8e","Type":"ContainerDied","Data":"f18874e6df55356bcb5e1583bd1408b2e4312f0f94eca9d54e04f8402497a275"} Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.429940 4745 scope.go:117] "RemoveContainer" containerID="b08c891927d1d9b43721ab5f99d80afc8bcdf78bcb7c87f13e350b82237d2fab" Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.450148 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9gtbx"] Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.459309 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9gtbx"] Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.472250 4745 scope.go:117] "RemoveContainer" containerID="e0a2bb8866b2a49bf35b23f2cf016df81bd1a86938a360ca03f8d591bfc6bbcd" Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.529777 4745 scope.go:117] "RemoveContainer" containerID="86ea154fc72c12ce2865c12f27e9cac327954dead185f22d2afda93851693cd2" Oct 10 14:04:39 crc kubenswrapper[4745]: E1010 14:04:39.534465 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86ea154fc72c12ce2865c12f27e9cac327954dead185f22d2afda93851693cd2\": container with ID starting with 86ea154fc72c12ce2865c12f27e9cac327954dead185f22d2afda93851693cd2 not found: ID does not exist" containerID="86ea154fc72c12ce2865c12f27e9cac327954dead185f22d2afda93851693cd2" Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.534526 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86ea154fc72c12ce2865c12f27e9cac327954dead185f22d2afda93851693cd2"} err="failed to get container status \"86ea154fc72c12ce2865c12f27e9cac327954dead185f22d2afda93851693cd2\": rpc error: code = NotFound desc = could not find container \"86ea154fc72c12ce2865c12f27e9cac327954dead185f22d2afda93851693cd2\": container with ID starting with 86ea154fc72c12ce2865c12f27e9cac327954dead185f22d2afda93851693cd2 not found: ID does not exist" Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.534563 4745 scope.go:117] "RemoveContainer" containerID="b08c891927d1d9b43721ab5f99d80afc8bcdf78bcb7c87f13e350b82237d2fab" Oct 10 14:04:39 crc kubenswrapper[4745]: E1010 14:04:39.535420 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b08c891927d1d9b43721ab5f99d80afc8bcdf78bcb7c87f13e350b82237d2fab\": container with ID starting with b08c891927d1d9b43721ab5f99d80afc8bcdf78bcb7c87f13e350b82237d2fab not found: ID does not exist" containerID="b08c891927d1d9b43721ab5f99d80afc8bcdf78bcb7c87f13e350b82237d2fab" Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.535463 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b08c891927d1d9b43721ab5f99d80afc8bcdf78bcb7c87f13e350b82237d2fab"} err="failed to get container status \"b08c891927d1d9b43721ab5f99d80afc8bcdf78bcb7c87f13e350b82237d2fab\": rpc error: code = NotFound desc = could not find container \"b08c891927d1d9b43721ab5f99d80afc8bcdf78bcb7c87f13e350b82237d2fab\": container with ID starting with b08c891927d1d9b43721ab5f99d80afc8bcdf78bcb7c87f13e350b82237d2fab not found: ID does not exist" Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.535494 4745 scope.go:117] "RemoveContainer" containerID="e0a2bb8866b2a49bf35b23f2cf016df81bd1a86938a360ca03f8d591bfc6bbcd" Oct 10 14:04:39 crc kubenswrapper[4745]: E1010 14:04:39.536208 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0a2bb8866b2a49bf35b23f2cf016df81bd1a86938a360ca03f8d591bfc6bbcd\": container with ID starting with e0a2bb8866b2a49bf35b23f2cf016df81bd1a86938a360ca03f8d591bfc6bbcd not found: ID does not exist" containerID="e0a2bb8866b2a49bf35b23f2cf016df81bd1a86938a360ca03f8d591bfc6bbcd" Oct 10 14:04:39 crc kubenswrapper[4745]: I1010 14:04:39.536241 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0a2bb8866b2a49bf35b23f2cf016df81bd1a86938a360ca03f8d591bfc6bbcd"} err="failed to get container status \"e0a2bb8866b2a49bf35b23f2cf016df81bd1a86938a360ca03f8d591bfc6bbcd\": rpc error: code = NotFound desc = could not find container \"e0a2bb8866b2a49bf35b23f2cf016df81bd1a86938a360ca03f8d591bfc6bbcd\": container with ID starting with e0a2bb8866b2a49bf35b23f2cf016df81bd1a86938a360ca03f8d591bfc6bbcd not found: ID does not exist" Oct 10 14:04:40 crc kubenswrapper[4745]: I1010 14:04:40.759419 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5e45037-8834-458d-b1e9-fa2b8fc7dc8e" path="/var/lib/kubelet/pods/b5e45037-8834-458d-b1e9-fa2b8fc7dc8e/volumes" Oct 10 14:04:46 crc kubenswrapper[4745]: I1010 14:04:46.187268 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:04:46 crc kubenswrapper[4745]: I1010 14:04:46.188029 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:05:16 crc kubenswrapper[4745]: I1010 14:05:16.187623 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:05:16 crc kubenswrapper[4745]: I1010 14:05:16.188137 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:05:46 crc kubenswrapper[4745]: I1010 14:05:46.186639 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:05:46 crc kubenswrapper[4745]: I1010 14:05:46.187235 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:05:46 crc kubenswrapper[4745]: I1010 14:05:46.187290 4745 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 14:05:46 crc kubenswrapper[4745]: I1010 14:05:46.188224 4745 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c"} pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 14:05:46 crc kubenswrapper[4745]: I1010 14:05:46.188287 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" containerID="cri-o://4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" gracePeriod=600 Oct 10 14:05:46 crc kubenswrapper[4745]: E1010 14:05:46.322775 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:05:47 crc kubenswrapper[4745]: I1010 14:05:47.129077 4745 generic.go:334] "Generic (PLEG): container finished" podID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" exitCode=0 Oct 10 14:05:47 crc kubenswrapper[4745]: I1010 14:05:47.129167 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerDied","Data":"4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c"} Oct 10 14:05:47 crc kubenswrapper[4745]: I1010 14:05:47.129242 4745 scope.go:117] "RemoveContainer" containerID="1f503d509545908a6f3ef6fd56dd209acab0d1ff66704707b763337227936e6d" Oct 10 14:05:47 crc kubenswrapper[4745]: I1010 14:05:47.130247 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:05:47 crc kubenswrapper[4745]: E1010 14:05:47.130860 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:06:00 crc kubenswrapper[4745]: I1010 14:06:00.744912 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:06:00 crc kubenswrapper[4745]: E1010 14:06:00.745680 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:06:15 crc kubenswrapper[4745]: I1010 14:06:15.745538 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:06:15 crc kubenswrapper[4745]: E1010 14:06:15.746419 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:06:30 crc kubenswrapper[4745]: I1010 14:06:30.745936 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:06:30 crc kubenswrapper[4745]: E1010 14:06:30.747150 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:06:42 crc kubenswrapper[4745]: I1010 14:06:42.745766 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:06:42 crc kubenswrapper[4745]: E1010 14:06:42.747072 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:06:53 crc kubenswrapper[4745]: I1010 14:06:53.745107 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:06:53 crc kubenswrapper[4745]: E1010 14:06:53.745908 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:07:05 crc kubenswrapper[4745]: I1010 14:07:05.745477 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:07:05 crc kubenswrapper[4745]: E1010 14:07:05.746437 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:07:17 crc kubenswrapper[4745]: I1010 14:07:17.745480 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:07:17 crc kubenswrapper[4745]: E1010 14:07:17.746300 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:07:32 crc kubenswrapper[4745]: I1010 14:07:32.744959 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:07:32 crc kubenswrapper[4745]: E1010 14:07:32.745634 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:07:45 crc kubenswrapper[4745]: I1010 14:07:45.747143 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:07:45 crc kubenswrapper[4745]: E1010 14:07:45.748171 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:07:57 crc kubenswrapper[4745]: I1010 14:07:57.745845 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:07:57 crc kubenswrapper[4745]: E1010 14:07:57.746838 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:08:10 crc kubenswrapper[4745]: I1010 14:08:10.745975 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:08:10 crc kubenswrapper[4745]: E1010 14:08:10.746843 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:08:21 crc kubenswrapper[4745]: I1010 14:08:21.745197 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:08:21 crc kubenswrapper[4745]: E1010 14:08:21.745835 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:08:34 crc kubenswrapper[4745]: I1010 14:08:34.744960 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:08:34 crc kubenswrapper[4745]: E1010 14:08:34.745676 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:08:47 crc kubenswrapper[4745]: I1010 14:08:47.744988 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:08:47 crc kubenswrapper[4745]: E1010 14:08:47.745889 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:09:00 crc kubenswrapper[4745]: I1010 14:09:00.746491 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:09:00 crc kubenswrapper[4745]: E1010 14:09:00.747666 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:09:14 crc kubenswrapper[4745]: I1010 14:09:14.744911 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:09:14 crc kubenswrapper[4745]: E1010 14:09:14.745813 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:09:27 crc kubenswrapper[4745]: I1010 14:09:27.745356 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:09:27 crc kubenswrapper[4745]: E1010 14:09:27.746159 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:09:38 crc kubenswrapper[4745]: I1010 14:09:38.744860 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:09:38 crc kubenswrapper[4745]: E1010 14:09:38.745802 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:09:53 crc kubenswrapper[4745]: I1010 14:09:53.745181 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:09:53 crc kubenswrapper[4745]: E1010 14:09:53.745808 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:10:06 crc kubenswrapper[4745]: I1010 14:10:06.758695 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:10:06 crc kubenswrapper[4745]: E1010 14:10:06.759988 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:10:19 crc kubenswrapper[4745]: I1010 14:10:19.745313 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:10:19 crc kubenswrapper[4745]: E1010 14:10:19.746243 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:10:34 crc kubenswrapper[4745]: I1010 14:10:34.746288 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:10:34 crc kubenswrapper[4745]: E1010 14:10:34.747462 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:10:46 crc kubenswrapper[4745]: I1010 14:10:46.757008 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:10:47 crc kubenswrapper[4745]: I1010 14:10:47.178605 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"c465e4549e850aec30db19773f9543087be6ef4e34f1ddf2ef454efa7ce84172"} Oct 10 14:12:20 crc kubenswrapper[4745]: I1010 14:12:20.822974 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jdr5w"] Oct 10 14:12:20 crc kubenswrapper[4745]: E1010 14:12:20.824528 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5e45037-8834-458d-b1e9-fa2b8fc7dc8e" containerName="extract-content" Oct 10 14:12:20 crc kubenswrapper[4745]: I1010 14:12:20.824562 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5e45037-8834-458d-b1e9-fa2b8fc7dc8e" containerName="extract-content" Oct 10 14:12:20 crc kubenswrapper[4745]: E1010 14:12:20.824595 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5e45037-8834-458d-b1e9-fa2b8fc7dc8e" containerName="extract-utilities" Oct 10 14:12:20 crc kubenswrapper[4745]: I1010 14:12:20.824613 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5e45037-8834-458d-b1e9-fa2b8fc7dc8e" containerName="extract-utilities" Oct 10 14:12:20 crc kubenswrapper[4745]: E1010 14:12:20.824658 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5e45037-8834-458d-b1e9-fa2b8fc7dc8e" containerName="registry-server" Oct 10 14:12:20 crc kubenswrapper[4745]: I1010 14:12:20.824675 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5e45037-8834-458d-b1e9-fa2b8fc7dc8e" containerName="registry-server" Oct 10 14:12:20 crc kubenswrapper[4745]: I1010 14:12:20.827028 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5e45037-8834-458d-b1e9-fa2b8fc7dc8e" containerName="registry-server" Oct 10 14:12:20 crc kubenswrapper[4745]: I1010 14:12:20.829656 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdr5w" Oct 10 14:12:20 crc kubenswrapper[4745]: I1010 14:12:20.847160 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jdr5w"] Oct 10 14:12:20 crc kubenswrapper[4745]: I1010 14:12:20.968586 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/658a659e-7082-466a-99bf-842fa2aa28d8-catalog-content\") pod \"redhat-operators-jdr5w\" (UID: \"658a659e-7082-466a-99bf-842fa2aa28d8\") " pod="openshift-marketplace/redhat-operators-jdr5w" Oct 10 14:12:20 crc kubenswrapper[4745]: I1010 14:12:20.968831 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvvdp\" (UniqueName: \"kubernetes.io/projected/658a659e-7082-466a-99bf-842fa2aa28d8-kube-api-access-gvvdp\") pod \"redhat-operators-jdr5w\" (UID: \"658a659e-7082-466a-99bf-842fa2aa28d8\") " pod="openshift-marketplace/redhat-operators-jdr5w" Oct 10 14:12:20 crc kubenswrapper[4745]: I1010 14:12:20.969166 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/658a659e-7082-466a-99bf-842fa2aa28d8-utilities\") pod \"redhat-operators-jdr5w\" (UID: \"658a659e-7082-466a-99bf-842fa2aa28d8\") " pod="openshift-marketplace/redhat-operators-jdr5w" Oct 10 14:12:21 crc kubenswrapper[4745]: I1010 14:12:21.071825 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/658a659e-7082-466a-99bf-842fa2aa28d8-catalog-content\") pod \"redhat-operators-jdr5w\" (UID: \"658a659e-7082-466a-99bf-842fa2aa28d8\") " pod="openshift-marketplace/redhat-operators-jdr5w" Oct 10 14:12:21 crc kubenswrapper[4745]: I1010 14:12:21.071959 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvvdp\" (UniqueName: \"kubernetes.io/projected/658a659e-7082-466a-99bf-842fa2aa28d8-kube-api-access-gvvdp\") pod \"redhat-operators-jdr5w\" (UID: \"658a659e-7082-466a-99bf-842fa2aa28d8\") " pod="openshift-marketplace/redhat-operators-jdr5w" Oct 10 14:12:21 crc kubenswrapper[4745]: I1010 14:12:21.072118 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/658a659e-7082-466a-99bf-842fa2aa28d8-utilities\") pod \"redhat-operators-jdr5w\" (UID: \"658a659e-7082-466a-99bf-842fa2aa28d8\") " pod="openshift-marketplace/redhat-operators-jdr5w" Oct 10 14:12:21 crc kubenswrapper[4745]: I1010 14:12:21.072710 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/658a659e-7082-466a-99bf-842fa2aa28d8-catalog-content\") pod \"redhat-operators-jdr5w\" (UID: \"658a659e-7082-466a-99bf-842fa2aa28d8\") " pod="openshift-marketplace/redhat-operators-jdr5w" Oct 10 14:12:21 crc kubenswrapper[4745]: I1010 14:12:21.073130 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/658a659e-7082-466a-99bf-842fa2aa28d8-utilities\") pod \"redhat-operators-jdr5w\" (UID: \"658a659e-7082-466a-99bf-842fa2aa28d8\") " pod="openshift-marketplace/redhat-operators-jdr5w" Oct 10 14:12:21 crc kubenswrapper[4745]: I1010 14:12:21.097836 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvvdp\" (UniqueName: \"kubernetes.io/projected/658a659e-7082-466a-99bf-842fa2aa28d8-kube-api-access-gvvdp\") pod \"redhat-operators-jdr5w\" (UID: \"658a659e-7082-466a-99bf-842fa2aa28d8\") " pod="openshift-marketplace/redhat-operators-jdr5w" Oct 10 14:12:21 crc kubenswrapper[4745]: I1010 14:12:21.168373 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdr5w" Oct 10 14:12:21 crc kubenswrapper[4745]: I1010 14:12:21.691761 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jdr5w"] Oct 10 14:12:22 crc kubenswrapper[4745]: I1010 14:12:22.069358 4745 generic.go:334] "Generic (PLEG): container finished" podID="658a659e-7082-466a-99bf-842fa2aa28d8" containerID="2c2c889b1e863695305e617e32046369de310d78b7f668b4d69d584448c24764" exitCode=0 Oct 10 14:12:22 crc kubenswrapper[4745]: I1010 14:12:22.069405 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdr5w" event={"ID":"658a659e-7082-466a-99bf-842fa2aa28d8","Type":"ContainerDied","Data":"2c2c889b1e863695305e617e32046369de310d78b7f668b4d69d584448c24764"} Oct 10 14:12:22 crc kubenswrapper[4745]: I1010 14:12:22.070646 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdr5w" event={"ID":"658a659e-7082-466a-99bf-842fa2aa28d8","Type":"ContainerStarted","Data":"f8a0c2ec94e6458ffe61cdf5515b601588c5df3fc34ac6cdfd0e6cd822f81531"} Oct 10 14:12:22 crc kubenswrapper[4745]: I1010 14:12:22.077467 4745 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 14:12:23 crc kubenswrapper[4745]: I1010 14:12:23.099083 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdr5w" event={"ID":"658a659e-7082-466a-99bf-842fa2aa28d8","Type":"ContainerStarted","Data":"149ba951b0b50895a2857d436c192875c03c9d542bae53ed5ec5e36e10fefbe1"} Oct 10 14:12:24 crc kubenswrapper[4745]: I1010 14:12:24.114524 4745 generic.go:334] "Generic (PLEG): container finished" podID="658a659e-7082-466a-99bf-842fa2aa28d8" containerID="149ba951b0b50895a2857d436c192875c03c9d542bae53ed5ec5e36e10fefbe1" exitCode=0 Oct 10 14:12:24 crc kubenswrapper[4745]: I1010 14:12:24.114590 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdr5w" event={"ID":"658a659e-7082-466a-99bf-842fa2aa28d8","Type":"ContainerDied","Data":"149ba951b0b50895a2857d436c192875c03c9d542bae53ed5ec5e36e10fefbe1"} Oct 10 14:12:25 crc kubenswrapper[4745]: I1010 14:12:25.126918 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdr5w" event={"ID":"658a659e-7082-466a-99bf-842fa2aa28d8","Type":"ContainerStarted","Data":"9b5f55bb2ae027c8c026d76a5d76ad6737d2367a08455838f2ed67592a8f62c5"} Oct 10 14:12:25 crc kubenswrapper[4745]: I1010 14:12:25.157293 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jdr5w" podStartSLOduration=2.652400047 podStartE2EDuration="5.157273026s" podCreationTimestamp="2025-10-10 14:12:20 +0000 UTC" firstStartedPulling="2025-10-10 14:12:22.077242574 +0000 UTC m=+3255.974899337" lastFinishedPulling="2025-10-10 14:12:24.582115543 +0000 UTC m=+3258.479772316" observedRunningTime="2025-10-10 14:12:25.148308279 +0000 UTC m=+3259.045965052" watchObservedRunningTime="2025-10-10 14:12:25.157273026 +0000 UTC m=+3259.054929799" Oct 10 14:12:31 crc kubenswrapper[4745]: I1010 14:12:31.169144 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jdr5w" Oct 10 14:12:31 crc kubenswrapper[4745]: I1010 14:12:31.169810 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jdr5w" Oct 10 14:12:31 crc kubenswrapper[4745]: I1010 14:12:31.250520 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jdr5w" Oct 10 14:12:31 crc kubenswrapper[4745]: I1010 14:12:31.332468 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jdr5w" Oct 10 14:12:31 crc kubenswrapper[4745]: I1010 14:12:31.502031 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jdr5w"] Oct 10 14:12:33 crc kubenswrapper[4745]: I1010 14:12:33.206490 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jdr5w" podUID="658a659e-7082-466a-99bf-842fa2aa28d8" containerName="registry-server" containerID="cri-o://9b5f55bb2ae027c8c026d76a5d76ad6737d2367a08455838f2ed67592a8f62c5" gracePeriod=2 Oct 10 14:12:33 crc kubenswrapper[4745]: I1010 14:12:33.709416 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdr5w" Oct 10 14:12:33 crc kubenswrapper[4745]: I1010 14:12:33.843777 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/658a659e-7082-466a-99bf-842fa2aa28d8-catalog-content\") pod \"658a659e-7082-466a-99bf-842fa2aa28d8\" (UID: \"658a659e-7082-466a-99bf-842fa2aa28d8\") " Oct 10 14:12:33 crc kubenswrapper[4745]: I1010 14:12:33.843949 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/658a659e-7082-466a-99bf-842fa2aa28d8-utilities\") pod \"658a659e-7082-466a-99bf-842fa2aa28d8\" (UID: \"658a659e-7082-466a-99bf-842fa2aa28d8\") " Oct 10 14:12:33 crc kubenswrapper[4745]: I1010 14:12:33.843992 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvvdp\" (UniqueName: \"kubernetes.io/projected/658a659e-7082-466a-99bf-842fa2aa28d8-kube-api-access-gvvdp\") pod \"658a659e-7082-466a-99bf-842fa2aa28d8\" (UID: \"658a659e-7082-466a-99bf-842fa2aa28d8\") " Oct 10 14:12:33 crc kubenswrapper[4745]: I1010 14:12:33.845194 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/658a659e-7082-466a-99bf-842fa2aa28d8-utilities" (OuterVolumeSpecName: "utilities") pod "658a659e-7082-466a-99bf-842fa2aa28d8" (UID: "658a659e-7082-466a-99bf-842fa2aa28d8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:12:33 crc kubenswrapper[4745]: I1010 14:12:33.850288 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/658a659e-7082-466a-99bf-842fa2aa28d8-kube-api-access-gvvdp" (OuterVolumeSpecName: "kube-api-access-gvvdp") pod "658a659e-7082-466a-99bf-842fa2aa28d8" (UID: "658a659e-7082-466a-99bf-842fa2aa28d8"). InnerVolumeSpecName "kube-api-access-gvvdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:12:33 crc kubenswrapper[4745]: I1010 14:12:33.921337 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/658a659e-7082-466a-99bf-842fa2aa28d8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "658a659e-7082-466a-99bf-842fa2aa28d8" (UID: "658a659e-7082-466a-99bf-842fa2aa28d8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:12:33 crc kubenswrapper[4745]: I1010 14:12:33.946377 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/658a659e-7082-466a-99bf-842fa2aa28d8-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:12:33 crc kubenswrapper[4745]: I1010 14:12:33.946426 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvvdp\" (UniqueName: \"kubernetes.io/projected/658a659e-7082-466a-99bf-842fa2aa28d8-kube-api-access-gvvdp\") on node \"crc\" DevicePath \"\"" Oct 10 14:12:33 crc kubenswrapper[4745]: I1010 14:12:33.946447 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/658a659e-7082-466a-99bf-842fa2aa28d8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:12:34 crc kubenswrapper[4745]: I1010 14:12:34.221513 4745 generic.go:334] "Generic (PLEG): container finished" podID="658a659e-7082-466a-99bf-842fa2aa28d8" containerID="9b5f55bb2ae027c8c026d76a5d76ad6737d2367a08455838f2ed67592a8f62c5" exitCode=0 Oct 10 14:12:34 crc kubenswrapper[4745]: I1010 14:12:34.221604 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdr5w" Oct 10 14:12:34 crc kubenswrapper[4745]: I1010 14:12:34.221649 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdr5w" event={"ID":"658a659e-7082-466a-99bf-842fa2aa28d8","Type":"ContainerDied","Data":"9b5f55bb2ae027c8c026d76a5d76ad6737d2367a08455838f2ed67592a8f62c5"} Oct 10 14:12:34 crc kubenswrapper[4745]: I1010 14:12:34.222014 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdr5w" event={"ID":"658a659e-7082-466a-99bf-842fa2aa28d8","Type":"ContainerDied","Data":"f8a0c2ec94e6458ffe61cdf5515b601588c5df3fc34ac6cdfd0e6cd822f81531"} Oct 10 14:12:34 crc kubenswrapper[4745]: I1010 14:12:34.222071 4745 scope.go:117] "RemoveContainer" containerID="9b5f55bb2ae027c8c026d76a5d76ad6737d2367a08455838f2ed67592a8f62c5" Oct 10 14:12:34 crc kubenswrapper[4745]: I1010 14:12:34.263060 4745 scope.go:117] "RemoveContainer" containerID="149ba951b0b50895a2857d436c192875c03c9d542bae53ed5ec5e36e10fefbe1" Oct 10 14:12:34 crc kubenswrapper[4745]: I1010 14:12:34.270320 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jdr5w"] Oct 10 14:12:34 crc kubenswrapper[4745]: I1010 14:12:34.291851 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jdr5w"] Oct 10 14:12:34 crc kubenswrapper[4745]: I1010 14:12:34.304263 4745 scope.go:117] "RemoveContainer" containerID="2c2c889b1e863695305e617e32046369de310d78b7f668b4d69d584448c24764" Oct 10 14:12:34 crc kubenswrapper[4745]: I1010 14:12:34.349690 4745 scope.go:117] "RemoveContainer" containerID="9b5f55bb2ae027c8c026d76a5d76ad6737d2367a08455838f2ed67592a8f62c5" Oct 10 14:12:34 crc kubenswrapper[4745]: E1010 14:12:34.350199 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b5f55bb2ae027c8c026d76a5d76ad6737d2367a08455838f2ed67592a8f62c5\": container with ID starting with 9b5f55bb2ae027c8c026d76a5d76ad6737d2367a08455838f2ed67592a8f62c5 not found: ID does not exist" containerID="9b5f55bb2ae027c8c026d76a5d76ad6737d2367a08455838f2ed67592a8f62c5" Oct 10 14:12:34 crc kubenswrapper[4745]: I1010 14:12:34.350236 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b5f55bb2ae027c8c026d76a5d76ad6737d2367a08455838f2ed67592a8f62c5"} err="failed to get container status \"9b5f55bb2ae027c8c026d76a5d76ad6737d2367a08455838f2ed67592a8f62c5\": rpc error: code = NotFound desc = could not find container \"9b5f55bb2ae027c8c026d76a5d76ad6737d2367a08455838f2ed67592a8f62c5\": container with ID starting with 9b5f55bb2ae027c8c026d76a5d76ad6737d2367a08455838f2ed67592a8f62c5 not found: ID does not exist" Oct 10 14:12:34 crc kubenswrapper[4745]: I1010 14:12:34.350258 4745 scope.go:117] "RemoveContainer" containerID="149ba951b0b50895a2857d436c192875c03c9d542bae53ed5ec5e36e10fefbe1" Oct 10 14:12:34 crc kubenswrapper[4745]: E1010 14:12:34.350984 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"149ba951b0b50895a2857d436c192875c03c9d542bae53ed5ec5e36e10fefbe1\": container with ID starting with 149ba951b0b50895a2857d436c192875c03c9d542bae53ed5ec5e36e10fefbe1 not found: ID does not exist" containerID="149ba951b0b50895a2857d436c192875c03c9d542bae53ed5ec5e36e10fefbe1" Oct 10 14:12:34 crc kubenswrapper[4745]: I1010 14:12:34.351063 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"149ba951b0b50895a2857d436c192875c03c9d542bae53ed5ec5e36e10fefbe1"} err="failed to get container status \"149ba951b0b50895a2857d436c192875c03c9d542bae53ed5ec5e36e10fefbe1\": rpc error: code = NotFound desc = could not find container \"149ba951b0b50895a2857d436c192875c03c9d542bae53ed5ec5e36e10fefbe1\": container with ID starting with 149ba951b0b50895a2857d436c192875c03c9d542bae53ed5ec5e36e10fefbe1 not found: ID does not exist" Oct 10 14:12:34 crc kubenswrapper[4745]: I1010 14:12:34.351121 4745 scope.go:117] "RemoveContainer" containerID="2c2c889b1e863695305e617e32046369de310d78b7f668b4d69d584448c24764" Oct 10 14:12:34 crc kubenswrapper[4745]: E1010 14:12:34.351602 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c2c889b1e863695305e617e32046369de310d78b7f668b4d69d584448c24764\": container with ID starting with 2c2c889b1e863695305e617e32046369de310d78b7f668b4d69d584448c24764 not found: ID does not exist" containerID="2c2c889b1e863695305e617e32046369de310d78b7f668b4d69d584448c24764" Oct 10 14:12:34 crc kubenswrapper[4745]: I1010 14:12:34.351629 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c2c889b1e863695305e617e32046369de310d78b7f668b4d69d584448c24764"} err="failed to get container status \"2c2c889b1e863695305e617e32046369de310d78b7f668b4d69d584448c24764\": rpc error: code = NotFound desc = could not find container \"2c2c889b1e863695305e617e32046369de310d78b7f668b4d69d584448c24764\": container with ID starting with 2c2c889b1e863695305e617e32046369de310d78b7f668b4d69d584448c24764 not found: ID does not exist" Oct 10 14:12:34 crc kubenswrapper[4745]: I1010 14:12:34.761402 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="658a659e-7082-466a-99bf-842fa2aa28d8" path="/var/lib/kubelet/pods/658a659e-7082-466a-99bf-842fa2aa28d8/volumes" Oct 10 14:12:46 crc kubenswrapper[4745]: I1010 14:12:46.187700 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:12:46 crc kubenswrapper[4745]: I1010 14:12:46.188239 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:13:16 crc kubenswrapper[4745]: I1010 14:13:16.187361 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:13:16 crc kubenswrapper[4745]: I1010 14:13:16.187950 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:13:46 crc kubenswrapper[4745]: I1010 14:13:46.187013 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:13:46 crc kubenswrapper[4745]: I1010 14:13:46.187596 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:13:46 crc kubenswrapper[4745]: I1010 14:13:46.187640 4745 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 14:13:46 crc kubenswrapper[4745]: I1010 14:13:46.188348 4745 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c465e4549e850aec30db19773f9543087be6ef4e34f1ddf2ef454efa7ce84172"} pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 14:13:46 crc kubenswrapper[4745]: I1010 14:13:46.188418 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" containerID="cri-o://c465e4549e850aec30db19773f9543087be6ef4e34f1ddf2ef454efa7ce84172" gracePeriod=600 Oct 10 14:13:46 crc kubenswrapper[4745]: I1010 14:13:46.924354 4745 generic.go:334] "Generic (PLEG): container finished" podID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerID="c465e4549e850aec30db19773f9543087be6ef4e34f1ddf2ef454efa7ce84172" exitCode=0 Oct 10 14:13:46 crc kubenswrapper[4745]: I1010 14:13:46.924439 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerDied","Data":"c465e4549e850aec30db19773f9543087be6ef4e34f1ddf2ef454efa7ce84172"} Oct 10 14:13:46 crc kubenswrapper[4745]: I1010 14:13:46.924785 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264"} Oct 10 14:13:46 crc kubenswrapper[4745]: I1010 14:13:46.924809 4745 scope.go:117] "RemoveContainer" containerID="4dc7a101fd7ba3279d5c04e386941b5a711af8ac91da9bb46ae07dbb9a6aad6c" Oct 10 14:14:15 crc kubenswrapper[4745]: I1010 14:14:15.194116 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qscts"] Oct 10 14:14:15 crc kubenswrapper[4745]: E1010 14:14:15.195139 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="658a659e-7082-466a-99bf-842fa2aa28d8" containerName="extract-content" Oct 10 14:14:15 crc kubenswrapper[4745]: I1010 14:14:15.195151 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="658a659e-7082-466a-99bf-842fa2aa28d8" containerName="extract-content" Oct 10 14:14:15 crc kubenswrapper[4745]: E1010 14:14:15.195172 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="658a659e-7082-466a-99bf-842fa2aa28d8" containerName="extract-utilities" Oct 10 14:14:15 crc kubenswrapper[4745]: I1010 14:14:15.195178 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="658a659e-7082-466a-99bf-842fa2aa28d8" containerName="extract-utilities" Oct 10 14:14:15 crc kubenswrapper[4745]: E1010 14:14:15.195197 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="658a659e-7082-466a-99bf-842fa2aa28d8" containerName="registry-server" Oct 10 14:14:15 crc kubenswrapper[4745]: I1010 14:14:15.195203 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="658a659e-7082-466a-99bf-842fa2aa28d8" containerName="registry-server" Oct 10 14:14:15 crc kubenswrapper[4745]: I1010 14:14:15.195392 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="658a659e-7082-466a-99bf-842fa2aa28d8" containerName="registry-server" Oct 10 14:14:15 crc kubenswrapper[4745]: I1010 14:14:15.202063 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qscts" Oct 10 14:14:15 crc kubenswrapper[4745]: I1010 14:14:15.211412 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qscts"] Oct 10 14:14:15 crc kubenswrapper[4745]: I1010 14:14:15.337076 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8pbp\" (UniqueName: \"kubernetes.io/projected/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-kube-api-access-b8pbp\") pod \"certified-operators-qscts\" (UID: \"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45\") " pod="openshift-marketplace/certified-operators-qscts" Oct 10 14:14:15 crc kubenswrapper[4745]: I1010 14:14:15.337160 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-catalog-content\") pod \"certified-operators-qscts\" (UID: \"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45\") " pod="openshift-marketplace/certified-operators-qscts" Oct 10 14:14:15 crc kubenswrapper[4745]: I1010 14:14:15.337319 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-utilities\") pod \"certified-operators-qscts\" (UID: \"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45\") " pod="openshift-marketplace/certified-operators-qscts" Oct 10 14:14:15 crc kubenswrapper[4745]: I1010 14:14:15.439185 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-utilities\") pod \"certified-operators-qscts\" (UID: \"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45\") " pod="openshift-marketplace/certified-operators-qscts" Oct 10 14:14:15 crc kubenswrapper[4745]: I1010 14:14:15.439256 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8pbp\" (UniqueName: \"kubernetes.io/projected/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-kube-api-access-b8pbp\") pod \"certified-operators-qscts\" (UID: \"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45\") " pod="openshift-marketplace/certified-operators-qscts" Oct 10 14:14:15 crc kubenswrapper[4745]: I1010 14:14:15.439303 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-catalog-content\") pod \"certified-operators-qscts\" (UID: \"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45\") " pod="openshift-marketplace/certified-operators-qscts" Oct 10 14:14:15 crc kubenswrapper[4745]: I1010 14:14:15.439859 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-catalog-content\") pod \"certified-operators-qscts\" (UID: \"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45\") " pod="openshift-marketplace/certified-operators-qscts" Oct 10 14:14:15 crc kubenswrapper[4745]: I1010 14:14:15.440131 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-utilities\") pod \"certified-operators-qscts\" (UID: \"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45\") " pod="openshift-marketplace/certified-operators-qscts" Oct 10 14:14:15 crc kubenswrapper[4745]: I1010 14:14:15.467204 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8pbp\" (UniqueName: \"kubernetes.io/projected/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-kube-api-access-b8pbp\") pod \"certified-operators-qscts\" (UID: \"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45\") " pod="openshift-marketplace/certified-operators-qscts" Oct 10 14:14:15 crc kubenswrapper[4745]: I1010 14:14:15.537319 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qscts" Oct 10 14:14:16 crc kubenswrapper[4745]: I1010 14:14:16.032019 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qscts"] Oct 10 14:14:16 crc kubenswrapper[4745]: I1010 14:14:16.218808 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qscts" event={"ID":"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45","Type":"ContainerStarted","Data":"714cf20edad842e41c4f6ca40217e35efc1b069babdb82f88f7e89149a83759b"} Oct 10 14:14:17 crc kubenswrapper[4745]: I1010 14:14:17.230084 4745 generic.go:334] "Generic (PLEG): container finished" podID="2aeba7df-fa09-40c2-87c3-fe6ac9dffd45" containerID="6b2925f26a18babd54d6ae6c68452924ad5abf1bfb62308ec9612fc648aa2d7b" exitCode=0 Oct 10 14:14:17 crc kubenswrapper[4745]: I1010 14:14:17.230141 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qscts" event={"ID":"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45","Type":"ContainerDied","Data":"6b2925f26a18babd54d6ae6c68452924ad5abf1bfb62308ec9612fc648aa2d7b"} Oct 10 14:14:19 crc kubenswrapper[4745]: I1010 14:14:19.246705 4745 generic.go:334] "Generic (PLEG): container finished" podID="2aeba7df-fa09-40c2-87c3-fe6ac9dffd45" containerID="5fb561507a395ebfd3b0dc690a0a59c139b7b9295c60af06de26ea56460c7366" exitCode=0 Oct 10 14:14:19 crc kubenswrapper[4745]: I1010 14:14:19.246759 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qscts" event={"ID":"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45","Type":"ContainerDied","Data":"5fb561507a395ebfd3b0dc690a0a59c139b7b9295c60af06de26ea56460c7366"} Oct 10 14:14:20 crc kubenswrapper[4745]: I1010 14:14:20.269508 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qscts" event={"ID":"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45","Type":"ContainerStarted","Data":"4b8dcc4ea4eab355705d0eee122c5030a5d3e75fec146d4fe23fd0b3fdf01d79"} Oct 10 14:14:20 crc kubenswrapper[4745]: I1010 14:14:20.288633 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qscts" podStartSLOduration=2.8660968799999997 podStartE2EDuration="5.288607636s" podCreationTimestamp="2025-10-10 14:14:15 +0000 UTC" firstStartedPulling="2025-10-10 14:14:17.233020423 +0000 UTC m=+3371.130677196" lastFinishedPulling="2025-10-10 14:14:19.655531189 +0000 UTC m=+3373.553187952" observedRunningTime="2025-10-10 14:14:20.28601275 +0000 UTC m=+3374.183669513" watchObservedRunningTime="2025-10-10 14:14:20.288607636 +0000 UTC m=+3374.186264399" Oct 10 14:14:25 crc kubenswrapper[4745]: I1010 14:14:25.538173 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qscts" Oct 10 14:14:25 crc kubenswrapper[4745]: I1010 14:14:25.538638 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qscts" Oct 10 14:14:25 crc kubenswrapper[4745]: I1010 14:14:25.584377 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qscts" Oct 10 14:14:26 crc kubenswrapper[4745]: I1010 14:14:26.376485 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qscts" Oct 10 14:14:26 crc kubenswrapper[4745]: I1010 14:14:26.425435 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qscts"] Oct 10 14:14:28 crc kubenswrapper[4745]: I1010 14:14:28.340507 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qscts" podUID="2aeba7df-fa09-40c2-87c3-fe6ac9dffd45" containerName="registry-server" containerID="cri-o://4b8dcc4ea4eab355705d0eee122c5030a5d3e75fec146d4fe23fd0b3fdf01d79" gracePeriod=2 Oct 10 14:14:28 crc kubenswrapper[4745]: I1010 14:14:28.908337 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qscts" Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.030937 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-catalog-content\") pod \"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45\" (UID: \"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45\") " Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.031472 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-utilities\") pod \"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45\" (UID: \"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45\") " Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.031677 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8pbp\" (UniqueName: \"kubernetes.io/projected/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-kube-api-access-b8pbp\") pod \"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45\" (UID: \"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45\") " Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.032627 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-utilities" (OuterVolumeSpecName: "utilities") pod "2aeba7df-fa09-40c2-87c3-fe6ac9dffd45" (UID: "2aeba7df-fa09-40c2-87c3-fe6ac9dffd45"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.032880 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.037859 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-kube-api-access-b8pbp" (OuterVolumeSpecName: "kube-api-access-b8pbp") pod "2aeba7df-fa09-40c2-87c3-fe6ac9dffd45" (UID: "2aeba7df-fa09-40c2-87c3-fe6ac9dffd45"). InnerVolumeSpecName "kube-api-access-b8pbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.133979 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8pbp\" (UniqueName: \"kubernetes.io/projected/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-kube-api-access-b8pbp\") on node \"crc\" DevicePath \"\"" Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.334582 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2aeba7df-fa09-40c2-87c3-fe6ac9dffd45" (UID: "2aeba7df-fa09-40c2-87c3-fe6ac9dffd45"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.339592 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.349043 4745 generic.go:334] "Generic (PLEG): container finished" podID="2aeba7df-fa09-40c2-87c3-fe6ac9dffd45" containerID="4b8dcc4ea4eab355705d0eee122c5030a5d3e75fec146d4fe23fd0b3fdf01d79" exitCode=0 Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.349082 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qscts" Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.349151 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qscts" event={"ID":"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45","Type":"ContainerDied","Data":"4b8dcc4ea4eab355705d0eee122c5030a5d3e75fec146d4fe23fd0b3fdf01d79"} Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.349212 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qscts" event={"ID":"2aeba7df-fa09-40c2-87c3-fe6ac9dffd45","Type":"ContainerDied","Data":"714cf20edad842e41c4f6ca40217e35efc1b069babdb82f88f7e89149a83759b"} Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.349243 4745 scope.go:117] "RemoveContainer" containerID="4b8dcc4ea4eab355705d0eee122c5030a5d3e75fec146d4fe23fd0b3fdf01d79" Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.368258 4745 scope.go:117] "RemoveContainer" containerID="5fb561507a395ebfd3b0dc690a0a59c139b7b9295c60af06de26ea56460c7366" Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.389639 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qscts"] Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.402532 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qscts"] Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.412051 4745 scope.go:117] "RemoveContainer" containerID="6b2925f26a18babd54d6ae6c68452924ad5abf1bfb62308ec9612fc648aa2d7b" Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.578918 4745 scope.go:117] "RemoveContainer" containerID="4b8dcc4ea4eab355705d0eee122c5030a5d3e75fec146d4fe23fd0b3fdf01d79" Oct 10 14:14:29 crc kubenswrapper[4745]: E1010 14:14:29.580930 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b8dcc4ea4eab355705d0eee122c5030a5d3e75fec146d4fe23fd0b3fdf01d79\": container with ID starting with 4b8dcc4ea4eab355705d0eee122c5030a5d3e75fec146d4fe23fd0b3fdf01d79 not found: ID does not exist" containerID="4b8dcc4ea4eab355705d0eee122c5030a5d3e75fec146d4fe23fd0b3fdf01d79" Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.580978 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b8dcc4ea4eab355705d0eee122c5030a5d3e75fec146d4fe23fd0b3fdf01d79"} err="failed to get container status \"4b8dcc4ea4eab355705d0eee122c5030a5d3e75fec146d4fe23fd0b3fdf01d79\": rpc error: code = NotFound desc = could not find container \"4b8dcc4ea4eab355705d0eee122c5030a5d3e75fec146d4fe23fd0b3fdf01d79\": container with ID starting with 4b8dcc4ea4eab355705d0eee122c5030a5d3e75fec146d4fe23fd0b3fdf01d79 not found: ID does not exist" Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.581005 4745 scope.go:117] "RemoveContainer" containerID="5fb561507a395ebfd3b0dc690a0a59c139b7b9295c60af06de26ea56460c7366" Oct 10 14:14:29 crc kubenswrapper[4745]: E1010 14:14:29.584827 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fb561507a395ebfd3b0dc690a0a59c139b7b9295c60af06de26ea56460c7366\": container with ID starting with 5fb561507a395ebfd3b0dc690a0a59c139b7b9295c60af06de26ea56460c7366 not found: ID does not exist" containerID="5fb561507a395ebfd3b0dc690a0a59c139b7b9295c60af06de26ea56460c7366" Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.584874 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fb561507a395ebfd3b0dc690a0a59c139b7b9295c60af06de26ea56460c7366"} err="failed to get container status \"5fb561507a395ebfd3b0dc690a0a59c139b7b9295c60af06de26ea56460c7366\": rpc error: code = NotFound desc = could not find container \"5fb561507a395ebfd3b0dc690a0a59c139b7b9295c60af06de26ea56460c7366\": container with ID starting with 5fb561507a395ebfd3b0dc690a0a59c139b7b9295c60af06de26ea56460c7366 not found: ID does not exist" Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.584896 4745 scope.go:117] "RemoveContainer" containerID="6b2925f26a18babd54d6ae6c68452924ad5abf1bfb62308ec9612fc648aa2d7b" Oct 10 14:14:29 crc kubenswrapper[4745]: E1010 14:14:29.598876 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b2925f26a18babd54d6ae6c68452924ad5abf1bfb62308ec9612fc648aa2d7b\": container with ID starting with 6b2925f26a18babd54d6ae6c68452924ad5abf1bfb62308ec9612fc648aa2d7b not found: ID does not exist" containerID="6b2925f26a18babd54d6ae6c68452924ad5abf1bfb62308ec9612fc648aa2d7b" Oct 10 14:14:29 crc kubenswrapper[4745]: I1010 14:14:29.598919 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b2925f26a18babd54d6ae6c68452924ad5abf1bfb62308ec9612fc648aa2d7b"} err="failed to get container status \"6b2925f26a18babd54d6ae6c68452924ad5abf1bfb62308ec9612fc648aa2d7b\": rpc error: code = NotFound desc = could not find container \"6b2925f26a18babd54d6ae6c68452924ad5abf1bfb62308ec9612fc648aa2d7b\": container with ID starting with 6b2925f26a18babd54d6ae6c68452924ad5abf1bfb62308ec9612fc648aa2d7b not found: ID does not exist" Oct 10 14:14:30 crc kubenswrapper[4745]: I1010 14:14:30.766865 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aeba7df-fa09-40c2-87c3-fe6ac9dffd45" path="/var/lib/kubelet/pods/2aeba7df-fa09-40c2-87c3-fe6ac9dffd45/volumes" Oct 10 14:14:56 crc kubenswrapper[4745]: I1010 14:14:56.628628 4745 generic.go:334] "Generic (PLEG): container finished" podID="55dab8c7-31e2-4694-9f9e-a3272376cddd" containerID="3614647b235f87652bbfd4290e8f51f2c04d41c401b7a1827317a48b499da92f" exitCode=0 Oct 10 14:14:56 crc kubenswrapper[4745]: I1010 14:14:56.628786 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"55dab8c7-31e2-4694-9f9e-a3272376cddd","Type":"ContainerDied","Data":"3614647b235f87652bbfd4290e8f51f2c04d41c401b7a1827317a48b499da92f"} Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.045557 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.167797 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/55dab8c7-31e2-4694-9f9e-a3272376cddd-openstack-config\") pod \"55dab8c7-31e2-4694-9f9e-a3272376cddd\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.167895 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/55dab8c7-31e2-4694-9f9e-a3272376cddd-test-operator-ephemeral-temporary\") pod \"55dab8c7-31e2-4694-9f9e-a3272376cddd\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.167940 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-ca-certs\") pod \"55dab8c7-31e2-4694-9f9e-a3272376cddd\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.168000 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqpt4\" (UniqueName: \"kubernetes.io/projected/55dab8c7-31e2-4694-9f9e-a3272376cddd-kube-api-access-cqpt4\") pod \"55dab8c7-31e2-4694-9f9e-a3272376cddd\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.168130 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-openstack-config-secret\") pod \"55dab8c7-31e2-4694-9f9e-a3272376cddd\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.168185 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"55dab8c7-31e2-4694-9f9e-a3272376cddd\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.168218 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-ssh-key\") pod \"55dab8c7-31e2-4694-9f9e-a3272376cddd\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.168249 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/55dab8c7-31e2-4694-9f9e-a3272376cddd-test-operator-ephemeral-workdir\") pod \"55dab8c7-31e2-4694-9f9e-a3272376cddd\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.168303 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55dab8c7-31e2-4694-9f9e-a3272376cddd-config-data\") pod \"55dab8c7-31e2-4694-9f9e-a3272376cddd\" (UID: \"55dab8c7-31e2-4694-9f9e-a3272376cddd\") " Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.168674 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55dab8c7-31e2-4694-9f9e-a3272376cddd-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "55dab8c7-31e2-4694-9f9e-a3272376cddd" (UID: "55dab8c7-31e2-4694-9f9e-a3272376cddd"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.169397 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55dab8c7-31e2-4694-9f9e-a3272376cddd-config-data" (OuterVolumeSpecName: "config-data") pod "55dab8c7-31e2-4694-9f9e-a3272376cddd" (UID: "55dab8c7-31e2-4694-9f9e-a3272376cddd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.169431 4745 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/55dab8c7-31e2-4694-9f9e-a3272376cddd-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.174065 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55dab8c7-31e2-4694-9f9e-a3272376cddd-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "55dab8c7-31e2-4694-9f9e-a3272376cddd" (UID: "55dab8c7-31e2-4694-9f9e-a3272376cddd"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.177325 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55dab8c7-31e2-4694-9f9e-a3272376cddd-kube-api-access-cqpt4" (OuterVolumeSpecName: "kube-api-access-cqpt4") pod "55dab8c7-31e2-4694-9f9e-a3272376cddd" (UID: "55dab8c7-31e2-4694-9f9e-a3272376cddd"). InnerVolumeSpecName "kube-api-access-cqpt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.183478 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "test-operator-logs") pod "55dab8c7-31e2-4694-9f9e-a3272376cddd" (UID: "55dab8c7-31e2-4694-9f9e-a3272376cddd"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.197762 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "55dab8c7-31e2-4694-9f9e-a3272376cddd" (UID: "55dab8c7-31e2-4694-9f9e-a3272376cddd"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.208056 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "55dab8c7-31e2-4694-9f9e-a3272376cddd" (UID: "55dab8c7-31e2-4694-9f9e-a3272376cddd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.209767 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "55dab8c7-31e2-4694-9f9e-a3272376cddd" (UID: "55dab8c7-31e2-4694-9f9e-a3272376cddd"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.223307 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55dab8c7-31e2-4694-9f9e-a3272376cddd-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "55dab8c7-31e2-4694-9f9e-a3272376cddd" (UID: "55dab8c7-31e2-4694-9f9e-a3272376cddd"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.272015 4745 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/55dab8c7-31e2-4694-9f9e-a3272376cddd-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.272063 4745 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.272080 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqpt4\" (UniqueName: \"kubernetes.io/projected/55dab8c7-31e2-4694-9f9e-a3272376cddd-kube-api-access-cqpt4\") on node \"crc\" DevicePath \"\"" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.272098 4745 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.272144 4745 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.272161 4745 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55dab8c7-31e2-4694-9f9e-a3272376cddd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.272177 4745 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/55dab8c7-31e2-4694-9f9e-a3272376cddd-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.272192 4745 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55dab8c7-31e2-4694-9f9e-a3272376cddd-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.303290 4745 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.374500 4745 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.675890 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"55dab8c7-31e2-4694-9f9e-a3272376cddd","Type":"ContainerDied","Data":"22fe7f082cce9bd9583a7e8ee51ae1745f37fcb31e872d461c7497aa0c51e70e"} Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.675990 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22fe7f082cce9bd9583a7e8ee51ae1745f37fcb31e872d461c7497aa0c51e70e" Oct 10 14:14:58 crc kubenswrapper[4745]: I1010 14:14:58.676309 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.200253 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m"] Oct 10 14:15:00 crc kubenswrapper[4745]: E1010 14:15:00.200982 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55dab8c7-31e2-4694-9f9e-a3272376cddd" containerName="tempest-tests-tempest-tests-runner" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.201001 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="55dab8c7-31e2-4694-9f9e-a3272376cddd" containerName="tempest-tests-tempest-tests-runner" Oct 10 14:15:00 crc kubenswrapper[4745]: E1010 14:15:00.201043 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aeba7df-fa09-40c2-87c3-fe6ac9dffd45" containerName="extract-content" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.201051 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aeba7df-fa09-40c2-87c3-fe6ac9dffd45" containerName="extract-content" Oct 10 14:15:00 crc kubenswrapper[4745]: E1010 14:15:00.201066 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aeba7df-fa09-40c2-87c3-fe6ac9dffd45" containerName="extract-utilities" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.201075 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aeba7df-fa09-40c2-87c3-fe6ac9dffd45" containerName="extract-utilities" Oct 10 14:15:00 crc kubenswrapper[4745]: E1010 14:15:00.201095 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aeba7df-fa09-40c2-87c3-fe6ac9dffd45" containerName="registry-server" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.201103 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aeba7df-fa09-40c2-87c3-fe6ac9dffd45" containerName="registry-server" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.201340 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="55dab8c7-31e2-4694-9f9e-a3272376cddd" containerName="tempest-tests-tempest-tests-runner" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.201383 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aeba7df-fa09-40c2-87c3-fe6ac9dffd45" containerName="registry-server" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.202130 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.204982 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.205367 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.219846 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m"] Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.229904 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmcjp\" (UniqueName: \"kubernetes.io/projected/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-kube-api-access-dmcjp\") pod \"collect-profiles-29335095-2885m\" (UID: \"09bf50bd-d7d6-44aa-92c0-b616e78a34b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.230047 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-secret-volume\") pod \"collect-profiles-29335095-2885m\" (UID: \"09bf50bd-d7d6-44aa-92c0-b616e78a34b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.230155 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-config-volume\") pod \"collect-profiles-29335095-2885m\" (UID: \"09bf50bd-d7d6-44aa-92c0-b616e78a34b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.332283 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-config-volume\") pod \"collect-profiles-29335095-2885m\" (UID: \"09bf50bd-d7d6-44aa-92c0-b616e78a34b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.332454 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmcjp\" (UniqueName: \"kubernetes.io/projected/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-kube-api-access-dmcjp\") pod \"collect-profiles-29335095-2885m\" (UID: \"09bf50bd-d7d6-44aa-92c0-b616e78a34b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.332520 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-secret-volume\") pod \"collect-profiles-29335095-2885m\" (UID: \"09bf50bd-d7d6-44aa-92c0-b616e78a34b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.333565 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-config-volume\") pod \"collect-profiles-29335095-2885m\" (UID: \"09bf50bd-d7d6-44aa-92c0-b616e78a34b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.338807 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-secret-volume\") pod \"collect-profiles-29335095-2885m\" (UID: \"09bf50bd-d7d6-44aa-92c0-b616e78a34b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.347866 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmcjp\" (UniqueName: \"kubernetes.io/projected/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-kube-api-access-dmcjp\") pod \"collect-profiles-29335095-2885m\" (UID: \"09bf50bd-d7d6-44aa-92c0-b616e78a34b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.527757 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m" Oct 10 14:15:00 crc kubenswrapper[4745]: I1010 14:15:00.967478 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m"] Oct 10 14:15:00 crc kubenswrapper[4745]: W1010 14:15:00.978612 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09bf50bd_d7d6_44aa_92c0_b616e78a34b2.slice/crio-50e624fd3e2a795bd20cb76b8e53fd1be0ecd2400273f860da3c961be414e553 WatchSource:0}: Error finding container 50e624fd3e2a795bd20cb76b8e53fd1be0ecd2400273f860da3c961be414e553: Status 404 returned error can't find the container with id 50e624fd3e2a795bd20cb76b8e53fd1be0ecd2400273f860da3c961be414e553 Oct 10 14:15:01 crc kubenswrapper[4745]: I1010 14:15:01.710994 4745 generic.go:334] "Generic (PLEG): container finished" podID="09bf50bd-d7d6-44aa-92c0-b616e78a34b2" containerID="8a803db4aaeaea43773323af0196e5beffc5b00c10a33ed10e52e8103f2eae22" exitCode=0 Oct 10 14:15:01 crc kubenswrapper[4745]: I1010 14:15:01.711064 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m" event={"ID":"09bf50bd-d7d6-44aa-92c0-b616e78a34b2","Type":"ContainerDied","Data":"8a803db4aaeaea43773323af0196e5beffc5b00c10a33ed10e52e8103f2eae22"} Oct 10 14:15:01 crc kubenswrapper[4745]: I1010 14:15:01.711134 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m" event={"ID":"09bf50bd-d7d6-44aa-92c0-b616e78a34b2","Type":"ContainerStarted","Data":"50e624fd3e2a795bd20cb76b8e53fd1be0ecd2400273f860da3c961be414e553"} Oct 10 14:15:03 crc kubenswrapper[4745]: I1010 14:15:03.081951 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m" Oct 10 14:15:03 crc kubenswrapper[4745]: I1010 14:15:03.192913 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmcjp\" (UniqueName: \"kubernetes.io/projected/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-kube-api-access-dmcjp\") pod \"09bf50bd-d7d6-44aa-92c0-b616e78a34b2\" (UID: \"09bf50bd-d7d6-44aa-92c0-b616e78a34b2\") " Oct 10 14:15:03 crc kubenswrapper[4745]: I1010 14:15:03.193036 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-secret-volume\") pod \"09bf50bd-d7d6-44aa-92c0-b616e78a34b2\" (UID: \"09bf50bd-d7d6-44aa-92c0-b616e78a34b2\") " Oct 10 14:15:03 crc kubenswrapper[4745]: I1010 14:15:03.194433 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-config-volume\") pod \"09bf50bd-d7d6-44aa-92c0-b616e78a34b2\" (UID: \"09bf50bd-d7d6-44aa-92c0-b616e78a34b2\") " Oct 10 14:15:03 crc kubenswrapper[4745]: I1010 14:15:03.195527 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-config-volume" (OuterVolumeSpecName: "config-volume") pod "09bf50bd-d7d6-44aa-92c0-b616e78a34b2" (UID: "09bf50bd-d7d6-44aa-92c0-b616e78a34b2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:15:03 crc kubenswrapper[4745]: I1010 14:15:03.200014 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "09bf50bd-d7d6-44aa-92c0-b616e78a34b2" (UID: "09bf50bd-d7d6-44aa-92c0-b616e78a34b2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:15:03 crc kubenswrapper[4745]: I1010 14:15:03.200038 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-kube-api-access-dmcjp" (OuterVolumeSpecName: "kube-api-access-dmcjp") pod "09bf50bd-d7d6-44aa-92c0-b616e78a34b2" (UID: "09bf50bd-d7d6-44aa-92c0-b616e78a34b2"). InnerVolumeSpecName "kube-api-access-dmcjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:15:03 crc kubenswrapper[4745]: I1010 14:15:03.297440 4745 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 14:15:03 crc kubenswrapper[4745]: I1010 14:15:03.297483 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmcjp\" (UniqueName: \"kubernetes.io/projected/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-kube-api-access-dmcjp\") on node \"crc\" DevicePath \"\"" Oct 10 14:15:03 crc kubenswrapper[4745]: I1010 14:15:03.297504 4745 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09bf50bd-d7d6-44aa-92c0-b616e78a34b2-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 14:15:03 crc kubenswrapper[4745]: I1010 14:15:03.734608 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m" event={"ID":"09bf50bd-d7d6-44aa-92c0-b616e78a34b2","Type":"ContainerDied","Data":"50e624fd3e2a795bd20cb76b8e53fd1be0ecd2400273f860da3c961be414e553"} Oct 10 14:15:03 crc kubenswrapper[4745]: I1010 14:15:03.734659 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50e624fd3e2a795bd20cb76b8e53fd1be0ecd2400273f860da3c961be414e553" Oct 10 14:15:03 crc kubenswrapper[4745]: I1010 14:15:03.734712 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335095-2885m" Oct 10 14:15:04 crc kubenswrapper[4745]: I1010 14:15:04.162011 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5"] Oct 10 14:15:04 crc kubenswrapper[4745]: I1010 14:15:04.172612 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335050-ds4l5"] Oct 10 14:15:04 crc kubenswrapper[4745]: I1010 14:15:04.759283 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eed511e-b79b-47ee-88ec-4fa9fb65146e" path="/var/lib/kubelet/pods/9eed511e-b79b-47ee-88ec-4fa9fb65146e/volumes" Oct 10 14:15:10 crc kubenswrapper[4745]: I1010 14:15:10.778171 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 10 14:15:10 crc kubenswrapper[4745]: E1010 14:15:10.780204 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09bf50bd-d7d6-44aa-92c0-b616e78a34b2" containerName="collect-profiles" Oct 10 14:15:10 crc kubenswrapper[4745]: I1010 14:15:10.780233 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="09bf50bd-d7d6-44aa-92c0-b616e78a34b2" containerName="collect-profiles" Oct 10 14:15:10 crc kubenswrapper[4745]: I1010 14:15:10.780986 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="09bf50bd-d7d6-44aa-92c0-b616e78a34b2" containerName="collect-profiles" Oct 10 14:15:10 crc kubenswrapper[4745]: I1010 14:15:10.782195 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 14:15:10 crc kubenswrapper[4745]: I1010 14:15:10.789940 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 10 14:15:10 crc kubenswrapper[4745]: I1010 14:15:10.795126 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-7q7m2" Oct 10 14:15:10 crc kubenswrapper[4745]: I1010 14:15:10.953346 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"cb7b00b8-301c-4b88-a189-2019e53366ec\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 14:15:10 crc kubenswrapper[4745]: I1010 14:15:10.953936 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w7g8\" (UniqueName: \"kubernetes.io/projected/cb7b00b8-301c-4b88-a189-2019e53366ec-kube-api-access-4w7g8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"cb7b00b8-301c-4b88-a189-2019e53366ec\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 14:15:11 crc kubenswrapper[4745]: I1010 14:15:11.055670 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w7g8\" (UniqueName: \"kubernetes.io/projected/cb7b00b8-301c-4b88-a189-2019e53366ec-kube-api-access-4w7g8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"cb7b00b8-301c-4b88-a189-2019e53366ec\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 14:15:11 crc kubenswrapper[4745]: I1010 14:15:11.055792 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"cb7b00b8-301c-4b88-a189-2019e53366ec\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 14:15:11 crc kubenswrapper[4745]: I1010 14:15:11.056268 4745 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"cb7b00b8-301c-4b88-a189-2019e53366ec\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 14:15:11 crc kubenswrapper[4745]: I1010 14:15:11.082357 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w7g8\" (UniqueName: \"kubernetes.io/projected/cb7b00b8-301c-4b88-a189-2019e53366ec-kube-api-access-4w7g8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"cb7b00b8-301c-4b88-a189-2019e53366ec\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 14:15:11 crc kubenswrapper[4745]: I1010 14:15:11.088306 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"cb7b00b8-301c-4b88-a189-2019e53366ec\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 14:15:11 crc kubenswrapper[4745]: I1010 14:15:11.142249 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 14:15:11 crc kubenswrapper[4745]: I1010 14:15:11.592776 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 10 14:15:11 crc kubenswrapper[4745]: W1010 14:15:11.599064 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb7b00b8_301c_4b88_a189_2019e53366ec.slice/crio-49792ee0abf68478ebcf9b51e10498cd438185bade683a3375a4d447709bdf7d WatchSource:0}: Error finding container 49792ee0abf68478ebcf9b51e10498cd438185bade683a3375a4d447709bdf7d: Status 404 returned error can't find the container with id 49792ee0abf68478ebcf9b51e10498cd438185bade683a3375a4d447709bdf7d Oct 10 14:15:11 crc kubenswrapper[4745]: I1010 14:15:11.842624 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"cb7b00b8-301c-4b88-a189-2019e53366ec","Type":"ContainerStarted","Data":"49792ee0abf68478ebcf9b51e10498cd438185bade683a3375a4d447709bdf7d"} Oct 10 14:15:12 crc kubenswrapper[4745]: I1010 14:15:12.853501 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"cb7b00b8-301c-4b88-a189-2019e53366ec","Type":"ContainerStarted","Data":"c8d4200f36b53478536008b027f7fd57964a4a5f1b5df09ea17551b53ebc03e3"} Oct 10 14:15:12 crc kubenswrapper[4745]: I1010 14:15:12.869323 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.916704353 podStartE2EDuration="2.869302096s" podCreationTimestamp="2025-10-10 14:15:10 +0000 UTC" firstStartedPulling="2025-10-10 14:15:11.602638196 +0000 UTC m=+3425.500294969" lastFinishedPulling="2025-10-10 14:15:12.555235949 +0000 UTC m=+3426.452892712" observedRunningTime="2025-10-10 14:15:12.86630505 +0000 UTC m=+3426.763961863" watchObservedRunningTime="2025-10-10 14:15:12.869302096 +0000 UTC m=+3426.766958869" Oct 10 14:15:22 crc kubenswrapper[4745]: I1010 14:15:22.926985 4745 scope.go:117] "RemoveContainer" containerID="293b6ca4af800b62d70c93b79a536134a5f6b42a1aad6c0db460acfdd3e51754" Oct 10 14:15:30 crc kubenswrapper[4745]: I1010 14:15:30.231601 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2mb5t/must-gather-rkp6m"] Oct 10 14:15:30 crc kubenswrapper[4745]: I1010 14:15:30.233655 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2mb5t/must-gather-rkp6m" Oct 10 14:15:30 crc kubenswrapper[4745]: I1010 14:15:30.242030 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-2mb5t"/"default-dockercfg-gvxzx" Oct 10 14:15:30 crc kubenswrapper[4745]: I1010 14:15:30.242284 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2mb5t/must-gather-rkp6m"] Oct 10 14:15:30 crc kubenswrapper[4745]: I1010 14:15:30.242388 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2mb5t"/"openshift-service-ca.crt" Oct 10 14:15:30 crc kubenswrapper[4745]: I1010 14:15:30.242679 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2mb5t"/"kube-root-ca.crt" Oct 10 14:15:30 crc kubenswrapper[4745]: I1010 14:15:30.417684 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/59d60977-5951-405c-ac97-7c9d6dfd68f5-must-gather-output\") pod \"must-gather-rkp6m\" (UID: \"59d60977-5951-405c-ac97-7c9d6dfd68f5\") " pod="openshift-must-gather-2mb5t/must-gather-rkp6m" Oct 10 14:15:30 crc kubenswrapper[4745]: I1010 14:15:30.418172 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sql7p\" (UniqueName: \"kubernetes.io/projected/59d60977-5951-405c-ac97-7c9d6dfd68f5-kube-api-access-sql7p\") pod \"must-gather-rkp6m\" (UID: \"59d60977-5951-405c-ac97-7c9d6dfd68f5\") " pod="openshift-must-gather-2mb5t/must-gather-rkp6m" Oct 10 14:15:30 crc kubenswrapper[4745]: I1010 14:15:30.520267 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/59d60977-5951-405c-ac97-7c9d6dfd68f5-must-gather-output\") pod \"must-gather-rkp6m\" (UID: \"59d60977-5951-405c-ac97-7c9d6dfd68f5\") " pod="openshift-must-gather-2mb5t/must-gather-rkp6m" Oct 10 14:15:30 crc kubenswrapper[4745]: I1010 14:15:30.520511 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sql7p\" (UniqueName: \"kubernetes.io/projected/59d60977-5951-405c-ac97-7c9d6dfd68f5-kube-api-access-sql7p\") pod \"must-gather-rkp6m\" (UID: \"59d60977-5951-405c-ac97-7c9d6dfd68f5\") " pod="openshift-must-gather-2mb5t/must-gather-rkp6m" Oct 10 14:15:30 crc kubenswrapper[4745]: I1010 14:15:30.521142 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/59d60977-5951-405c-ac97-7c9d6dfd68f5-must-gather-output\") pod \"must-gather-rkp6m\" (UID: \"59d60977-5951-405c-ac97-7c9d6dfd68f5\") " pod="openshift-must-gather-2mb5t/must-gather-rkp6m" Oct 10 14:15:30 crc kubenswrapper[4745]: I1010 14:15:30.539359 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sql7p\" (UniqueName: \"kubernetes.io/projected/59d60977-5951-405c-ac97-7c9d6dfd68f5-kube-api-access-sql7p\") pod \"must-gather-rkp6m\" (UID: \"59d60977-5951-405c-ac97-7c9d6dfd68f5\") " pod="openshift-must-gather-2mb5t/must-gather-rkp6m" Oct 10 14:15:30 crc kubenswrapper[4745]: I1010 14:15:30.564271 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2mb5t/must-gather-rkp6m" Oct 10 14:15:31 crc kubenswrapper[4745]: I1010 14:15:31.021913 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2mb5t/must-gather-rkp6m"] Oct 10 14:15:31 crc kubenswrapper[4745]: I1010 14:15:31.084435 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2mb5t/must-gather-rkp6m" event={"ID":"59d60977-5951-405c-ac97-7c9d6dfd68f5","Type":"ContainerStarted","Data":"a912e15f416e2f58cd13dd65a0e58f97bb443f092ad12184e4393dc6f198aa2d"} Oct 10 14:15:37 crc kubenswrapper[4745]: I1010 14:15:37.139796 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2mb5t/must-gather-rkp6m" event={"ID":"59d60977-5951-405c-ac97-7c9d6dfd68f5","Type":"ContainerStarted","Data":"a2171f4b84612b69d263d5b3c0bce6c81aaa9378cb05a01b7a4a071174eb5183"} Oct 10 14:15:37 crc kubenswrapper[4745]: I1010 14:15:37.140298 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2mb5t/must-gather-rkp6m" event={"ID":"59d60977-5951-405c-ac97-7c9d6dfd68f5","Type":"ContainerStarted","Data":"d2211773dcc78db2c4a4cc8dd94a0b34073237015e0021d0825c6cecf525ac9c"} Oct 10 14:15:40 crc kubenswrapper[4745]: I1010 14:15:40.167614 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2mb5t/must-gather-rkp6m" podStartSLOduration=4.566208553 podStartE2EDuration="10.167593321s" podCreationTimestamp="2025-10-10 14:15:30 +0000 UTC" firstStartedPulling="2025-10-10 14:15:31.019580202 +0000 UTC m=+3444.917236965" lastFinishedPulling="2025-10-10 14:15:36.62096497 +0000 UTC m=+3450.518621733" observedRunningTime="2025-10-10 14:15:37.160133942 +0000 UTC m=+3451.057790705" watchObservedRunningTime="2025-10-10 14:15:40.167593321 +0000 UTC m=+3454.065250084" Oct 10 14:15:40 crc kubenswrapper[4745]: I1010 14:15:40.170647 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2mb5t/crc-debug-jcps2"] Oct 10 14:15:40 crc kubenswrapper[4745]: I1010 14:15:40.172125 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2mb5t/crc-debug-jcps2" Oct 10 14:15:40 crc kubenswrapper[4745]: I1010 14:15:40.308032 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t88tv\" (UniqueName: \"kubernetes.io/projected/5a7e2ddc-4584-43bf-8706-405a5f53b781-kube-api-access-t88tv\") pod \"crc-debug-jcps2\" (UID: \"5a7e2ddc-4584-43bf-8706-405a5f53b781\") " pod="openshift-must-gather-2mb5t/crc-debug-jcps2" Oct 10 14:15:40 crc kubenswrapper[4745]: I1010 14:15:40.308133 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5a7e2ddc-4584-43bf-8706-405a5f53b781-host\") pod \"crc-debug-jcps2\" (UID: \"5a7e2ddc-4584-43bf-8706-405a5f53b781\") " pod="openshift-must-gather-2mb5t/crc-debug-jcps2" Oct 10 14:15:40 crc kubenswrapper[4745]: I1010 14:15:40.410237 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t88tv\" (UniqueName: \"kubernetes.io/projected/5a7e2ddc-4584-43bf-8706-405a5f53b781-kube-api-access-t88tv\") pod \"crc-debug-jcps2\" (UID: \"5a7e2ddc-4584-43bf-8706-405a5f53b781\") " pod="openshift-must-gather-2mb5t/crc-debug-jcps2" Oct 10 14:15:40 crc kubenswrapper[4745]: I1010 14:15:40.410328 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5a7e2ddc-4584-43bf-8706-405a5f53b781-host\") pod \"crc-debug-jcps2\" (UID: \"5a7e2ddc-4584-43bf-8706-405a5f53b781\") " pod="openshift-must-gather-2mb5t/crc-debug-jcps2" Oct 10 14:15:40 crc kubenswrapper[4745]: I1010 14:15:40.410519 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5a7e2ddc-4584-43bf-8706-405a5f53b781-host\") pod \"crc-debug-jcps2\" (UID: \"5a7e2ddc-4584-43bf-8706-405a5f53b781\") " pod="openshift-must-gather-2mb5t/crc-debug-jcps2" Oct 10 14:15:40 crc kubenswrapper[4745]: I1010 14:15:40.445375 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t88tv\" (UniqueName: \"kubernetes.io/projected/5a7e2ddc-4584-43bf-8706-405a5f53b781-kube-api-access-t88tv\") pod \"crc-debug-jcps2\" (UID: \"5a7e2ddc-4584-43bf-8706-405a5f53b781\") " pod="openshift-must-gather-2mb5t/crc-debug-jcps2" Oct 10 14:15:40 crc kubenswrapper[4745]: I1010 14:15:40.494614 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2mb5t/crc-debug-jcps2" Oct 10 14:15:40 crc kubenswrapper[4745]: W1010 14:15:40.533871 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a7e2ddc_4584_43bf_8706_405a5f53b781.slice/crio-189b9c77a5af4dd0721c0d650a1e96a94cbd8b1e046ea2bbc3ed282c6ff97f9c WatchSource:0}: Error finding container 189b9c77a5af4dd0721c0d650a1e96a94cbd8b1e046ea2bbc3ed282c6ff97f9c: Status 404 returned error can't find the container with id 189b9c77a5af4dd0721c0d650a1e96a94cbd8b1e046ea2bbc3ed282c6ff97f9c Oct 10 14:15:41 crc kubenswrapper[4745]: I1010 14:15:41.174544 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2mb5t/crc-debug-jcps2" event={"ID":"5a7e2ddc-4584-43bf-8706-405a5f53b781","Type":"ContainerStarted","Data":"189b9c77a5af4dd0721c0d650a1e96a94cbd8b1e046ea2bbc3ed282c6ff97f9c"} Oct 10 14:15:42 crc kubenswrapper[4745]: I1010 14:15:42.486579 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gxll9"] Oct 10 14:15:42 crc kubenswrapper[4745]: I1010 14:15:42.488857 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxll9" Oct 10 14:15:42 crc kubenswrapper[4745]: I1010 14:15:42.510764 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gxll9"] Oct 10 14:15:42 crc kubenswrapper[4745]: I1010 14:15:42.558796 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e33f867-88af-4331-b25f-1ac55b138e4d-utilities\") pod \"community-operators-gxll9\" (UID: \"7e33f867-88af-4331-b25f-1ac55b138e4d\") " pod="openshift-marketplace/community-operators-gxll9" Oct 10 14:15:42 crc kubenswrapper[4745]: I1010 14:15:42.558883 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e33f867-88af-4331-b25f-1ac55b138e4d-catalog-content\") pod \"community-operators-gxll9\" (UID: \"7e33f867-88af-4331-b25f-1ac55b138e4d\") " pod="openshift-marketplace/community-operators-gxll9" Oct 10 14:15:42 crc kubenswrapper[4745]: I1010 14:15:42.559685 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z94nc\" (UniqueName: \"kubernetes.io/projected/7e33f867-88af-4331-b25f-1ac55b138e4d-kube-api-access-z94nc\") pod \"community-operators-gxll9\" (UID: \"7e33f867-88af-4331-b25f-1ac55b138e4d\") " pod="openshift-marketplace/community-operators-gxll9" Oct 10 14:15:42 crc kubenswrapper[4745]: I1010 14:15:42.662201 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e33f867-88af-4331-b25f-1ac55b138e4d-utilities\") pod \"community-operators-gxll9\" (UID: \"7e33f867-88af-4331-b25f-1ac55b138e4d\") " pod="openshift-marketplace/community-operators-gxll9" Oct 10 14:15:42 crc kubenswrapper[4745]: I1010 14:15:42.662322 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e33f867-88af-4331-b25f-1ac55b138e4d-catalog-content\") pod \"community-operators-gxll9\" (UID: \"7e33f867-88af-4331-b25f-1ac55b138e4d\") " pod="openshift-marketplace/community-operators-gxll9" Oct 10 14:15:42 crc kubenswrapper[4745]: I1010 14:15:42.662401 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z94nc\" (UniqueName: \"kubernetes.io/projected/7e33f867-88af-4331-b25f-1ac55b138e4d-kube-api-access-z94nc\") pod \"community-operators-gxll9\" (UID: \"7e33f867-88af-4331-b25f-1ac55b138e4d\") " pod="openshift-marketplace/community-operators-gxll9" Oct 10 14:15:42 crc kubenswrapper[4745]: I1010 14:15:42.663305 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e33f867-88af-4331-b25f-1ac55b138e4d-utilities\") pod \"community-operators-gxll9\" (UID: \"7e33f867-88af-4331-b25f-1ac55b138e4d\") " pod="openshift-marketplace/community-operators-gxll9" Oct 10 14:15:42 crc kubenswrapper[4745]: I1010 14:15:42.663314 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e33f867-88af-4331-b25f-1ac55b138e4d-catalog-content\") pod \"community-operators-gxll9\" (UID: \"7e33f867-88af-4331-b25f-1ac55b138e4d\") " pod="openshift-marketplace/community-operators-gxll9" Oct 10 14:15:42 crc kubenswrapper[4745]: I1010 14:15:42.683765 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z94nc\" (UniqueName: \"kubernetes.io/projected/7e33f867-88af-4331-b25f-1ac55b138e4d-kube-api-access-z94nc\") pod \"community-operators-gxll9\" (UID: \"7e33f867-88af-4331-b25f-1ac55b138e4d\") " pod="openshift-marketplace/community-operators-gxll9" Oct 10 14:15:42 crc kubenswrapper[4745]: I1010 14:15:42.811369 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxll9" Oct 10 14:15:43 crc kubenswrapper[4745]: I1010 14:15:43.354498 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gxll9"] Oct 10 14:15:44 crc kubenswrapper[4745]: I1010 14:15:44.203807 4745 generic.go:334] "Generic (PLEG): container finished" podID="7e33f867-88af-4331-b25f-1ac55b138e4d" containerID="846b3e14ee3d5f0eac51c86044e21222ade366a32c019dc503d37165a621fe4a" exitCode=0 Oct 10 14:15:44 crc kubenswrapper[4745]: I1010 14:15:44.203864 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxll9" event={"ID":"7e33f867-88af-4331-b25f-1ac55b138e4d","Type":"ContainerDied","Data":"846b3e14ee3d5f0eac51c86044e21222ade366a32c019dc503d37165a621fe4a"} Oct 10 14:15:44 crc kubenswrapper[4745]: I1010 14:15:44.204220 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxll9" event={"ID":"7e33f867-88af-4331-b25f-1ac55b138e4d","Type":"ContainerStarted","Data":"4c4cb26802b5393053b081c3b07a1c535c426cf49451bc9940ea4c2b7e72541d"} Oct 10 14:15:45 crc kubenswrapper[4745]: I1010 14:15:45.214328 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxll9" event={"ID":"7e33f867-88af-4331-b25f-1ac55b138e4d","Type":"ContainerStarted","Data":"22fa3e6523fa81b5983fbda2fb5eea8e2e9887744d8a398e91581629ff2171cb"} Oct 10 14:15:46 crc kubenswrapper[4745]: I1010 14:15:46.186452 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:15:46 crc kubenswrapper[4745]: I1010 14:15:46.186705 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:15:46 crc kubenswrapper[4745]: I1010 14:15:46.227142 4745 generic.go:334] "Generic (PLEG): container finished" podID="7e33f867-88af-4331-b25f-1ac55b138e4d" containerID="22fa3e6523fa81b5983fbda2fb5eea8e2e9887744d8a398e91581629ff2171cb" exitCode=0 Oct 10 14:15:46 crc kubenswrapper[4745]: I1010 14:15:46.227191 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxll9" event={"ID":"7e33f867-88af-4331-b25f-1ac55b138e4d","Type":"ContainerDied","Data":"22fa3e6523fa81b5983fbda2fb5eea8e2e9887744d8a398e91581629ff2171cb"} Oct 10 14:15:53 crc kubenswrapper[4745]: I1010 14:15:53.307585 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxll9" event={"ID":"7e33f867-88af-4331-b25f-1ac55b138e4d","Type":"ContainerStarted","Data":"233cde53a5c42db7a3bbdf0124d0e644398dbf744735a5a7b27d0f605e59ccba"} Oct 10 14:15:53 crc kubenswrapper[4745]: I1010 14:15:53.309019 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2mb5t/crc-debug-jcps2" event={"ID":"5a7e2ddc-4584-43bf-8706-405a5f53b781","Type":"ContainerStarted","Data":"52dfac7e7ddc8facd62fb7de45ce63d37600f743df740eb3a4d7505170bdac35"} Oct 10 14:15:53 crc kubenswrapper[4745]: I1010 14:15:53.364788 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gxll9" podStartSLOduration=2.879317313 podStartE2EDuration="11.364759928s" podCreationTimestamp="2025-10-10 14:15:42 +0000 UTC" firstStartedPulling="2025-10-10 14:15:44.205702468 +0000 UTC m=+3458.103359231" lastFinishedPulling="2025-10-10 14:15:52.691145073 +0000 UTC m=+3466.588801846" observedRunningTime="2025-10-10 14:15:53.347258145 +0000 UTC m=+3467.244914908" watchObservedRunningTime="2025-10-10 14:15:53.364759928 +0000 UTC m=+3467.262416691" Oct 10 14:15:53 crc kubenswrapper[4745]: I1010 14:15:53.376697 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2mb5t/crc-debug-jcps2" podStartSLOduration=1.771110913 podStartE2EDuration="13.37667743s" podCreationTimestamp="2025-10-10 14:15:40 +0000 UTC" firstStartedPulling="2025-10-10 14:15:40.535878527 +0000 UTC m=+3454.433535290" lastFinishedPulling="2025-10-10 14:15:52.141445054 +0000 UTC m=+3466.039101807" observedRunningTime="2025-10-10 14:15:53.371066898 +0000 UTC m=+3467.268723661" watchObservedRunningTime="2025-10-10 14:15:53.37667743 +0000 UTC m=+3467.274334193" Oct 10 14:16:02 crc kubenswrapper[4745]: I1010 14:16:02.812449 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gxll9" Oct 10 14:16:02 crc kubenswrapper[4745]: I1010 14:16:02.813808 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gxll9" Oct 10 14:16:02 crc kubenswrapper[4745]: I1010 14:16:02.890241 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gxll9" Oct 10 14:16:03 crc kubenswrapper[4745]: I1010 14:16:03.450206 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gxll9" Oct 10 14:16:03 crc kubenswrapper[4745]: I1010 14:16:03.509519 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gxll9"] Oct 10 14:16:05 crc kubenswrapper[4745]: I1010 14:16:05.404535 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gxll9" podUID="7e33f867-88af-4331-b25f-1ac55b138e4d" containerName="registry-server" containerID="cri-o://233cde53a5c42db7a3bbdf0124d0e644398dbf744735a5a7b27d0f605e59ccba" gracePeriod=2 Oct 10 14:16:05 crc kubenswrapper[4745]: I1010 14:16:05.887340 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxll9" Oct 10 14:16:05 crc kubenswrapper[4745]: I1010 14:16:05.970395 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e33f867-88af-4331-b25f-1ac55b138e4d-utilities\") pod \"7e33f867-88af-4331-b25f-1ac55b138e4d\" (UID: \"7e33f867-88af-4331-b25f-1ac55b138e4d\") " Oct 10 14:16:05 crc kubenswrapper[4745]: I1010 14:16:05.970591 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e33f867-88af-4331-b25f-1ac55b138e4d-catalog-content\") pod \"7e33f867-88af-4331-b25f-1ac55b138e4d\" (UID: \"7e33f867-88af-4331-b25f-1ac55b138e4d\") " Oct 10 14:16:05 crc kubenswrapper[4745]: I1010 14:16:05.970613 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z94nc\" (UniqueName: \"kubernetes.io/projected/7e33f867-88af-4331-b25f-1ac55b138e4d-kube-api-access-z94nc\") pod \"7e33f867-88af-4331-b25f-1ac55b138e4d\" (UID: \"7e33f867-88af-4331-b25f-1ac55b138e4d\") " Oct 10 14:16:05 crc kubenswrapper[4745]: I1010 14:16:05.972517 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e33f867-88af-4331-b25f-1ac55b138e4d-utilities" (OuterVolumeSpecName: "utilities") pod "7e33f867-88af-4331-b25f-1ac55b138e4d" (UID: "7e33f867-88af-4331-b25f-1ac55b138e4d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:16:05 crc kubenswrapper[4745]: I1010 14:16:05.991783 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e33f867-88af-4331-b25f-1ac55b138e4d-kube-api-access-z94nc" (OuterVolumeSpecName: "kube-api-access-z94nc") pod "7e33f867-88af-4331-b25f-1ac55b138e4d" (UID: "7e33f867-88af-4331-b25f-1ac55b138e4d"). InnerVolumeSpecName "kube-api-access-z94nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.025829 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e33f867-88af-4331-b25f-1ac55b138e4d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e33f867-88af-4331-b25f-1ac55b138e4d" (UID: "7e33f867-88af-4331-b25f-1ac55b138e4d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.072747 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e33f867-88af-4331-b25f-1ac55b138e4d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.072779 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z94nc\" (UniqueName: \"kubernetes.io/projected/7e33f867-88af-4331-b25f-1ac55b138e4d-kube-api-access-z94nc\") on node \"crc\" DevicePath \"\"" Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.072790 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e33f867-88af-4331-b25f-1ac55b138e4d-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.418200 4745 generic.go:334] "Generic (PLEG): container finished" podID="7e33f867-88af-4331-b25f-1ac55b138e4d" containerID="233cde53a5c42db7a3bbdf0124d0e644398dbf744735a5a7b27d0f605e59ccba" exitCode=0 Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.418290 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxll9" Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.418288 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxll9" event={"ID":"7e33f867-88af-4331-b25f-1ac55b138e4d","Type":"ContainerDied","Data":"233cde53a5c42db7a3bbdf0124d0e644398dbf744735a5a7b27d0f605e59ccba"} Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.418602 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxll9" event={"ID":"7e33f867-88af-4331-b25f-1ac55b138e4d","Type":"ContainerDied","Data":"4c4cb26802b5393053b081c3b07a1c535c426cf49451bc9940ea4c2b7e72541d"} Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.418628 4745 scope.go:117] "RemoveContainer" containerID="233cde53a5c42db7a3bbdf0124d0e644398dbf744735a5a7b27d0f605e59ccba" Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.450128 4745 scope.go:117] "RemoveContainer" containerID="22fa3e6523fa81b5983fbda2fb5eea8e2e9887744d8a398e91581629ff2171cb" Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.470724 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gxll9"] Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.481255 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gxll9"] Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.487229 4745 scope.go:117] "RemoveContainer" containerID="846b3e14ee3d5f0eac51c86044e21222ade366a32c019dc503d37165a621fe4a" Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.521405 4745 scope.go:117] "RemoveContainer" containerID="233cde53a5c42db7a3bbdf0124d0e644398dbf744735a5a7b27d0f605e59ccba" Oct 10 14:16:06 crc kubenswrapper[4745]: E1010 14:16:06.521811 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"233cde53a5c42db7a3bbdf0124d0e644398dbf744735a5a7b27d0f605e59ccba\": container with ID starting with 233cde53a5c42db7a3bbdf0124d0e644398dbf744735a5a7b27d0f605e59ccba not found: ID does not exist" containerID="233cde53a5c42db7a3bbdf0124d0e644398dbf744735a5a7b27d0f605e59ccba" Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.521856 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"233cde53a5c42db7a3bbdf0124d0e644398dbf744735a5a7b27d0f605e59ccba"} err="failed to get container status \"233cde53a5c42db7a3bbdf0124d0e644398dbf744735a5a7b27d0f605e59ccba\": rpc error: code = NotFound desc = could not find container \"233cde53a5c42db7a3bbdf0124d0e644398dbf744735a5a7b27d0f605e59ccba\": container with ID starting with 233cde53a5c42db7a3bbdf0124d0e644398dbf744735a5a7b27d0f605e59ccba not found: ID does not exist" Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.521882 4745 scope.go:117] "RemoveContainer" containerID="22fa3e6523fa81b5983fbda2fb5eea8e2e9887744d8a398e91581629ff2171cb" Oct 10 14:16:06 crc kubenswrapper[4745]: E1010 14:16:06.522140 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22fa3e6523fa81b5983fbda2fb5eea8e2e9887744d8a398e91581629ff2171cb\": container with ID starting with 22fa3e6523fa81b5983fbda2fb5eea8e2e9887744d8a398e91581629ff2171cb not found: ID does not exist" containerID="22fa3e6523fa81b5983fbda2fb5eea8e2e9887744d8a398e91581629ff2171cb" Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.522163 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22fa3e6523fa81b5983fbda2fb5eea8e2e9887744d8a398e91581629ff2171cb"} err="failed to get container status \"22fa3e6523fa81b5983fbda2fb5eea8e2e9887744d8a398e91581629ff2171cb\": rpc error: code = NotFound desc = could not find container \"22fa3e6523fa81b5983fbda2fb5eea8e2e9887744d8a398e91581629ff2171cb\": container with ID starting with 22fa3e6523fa81b5983fbda2fb5eea8e2e9887744d8a398e91581629ff2171cb not found: ID does not exist" Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.522177 4745 scope.go:117] "RemoveContainer" containerID="846b3e14ee3d5f0eac51c86044e21222ade366a32c019dc503d37165a621fe4a" Oct 10 14:16:06 crc kubenswrapper[4745]: E1010 14:16:06.522547 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"846b3e14ee3d5f0eac51c86044e21222ade366a32c019dc503d37165a621fe4a\": container with ID starting with 846b3e14ee3d5f0eac51c86044e21222ade366a32c019dc503d37165a621fe4a not found: ID does not exist" containerID="846b3e14ee3d5f0eac51c86044e21222ade366a32c019dc503d37165a621fe4a" Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.522576 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"846b3e14ee3d5f0eac51c86044e21222ade366a32c019dc503d37165a621fe4a"} err="failed to get container status \"846b3e14ee3d5f0eac51c86044e21222ade366a32c019dc503d37165a621fe4a\": rpc error: code = NotFound desc = could not find container \"846b3e14ee3d5f0eac51c86044e21222ade366a32c019dc503d37165a621fe4a\": container with ID starting with 846b3e14ee3d5f0eac51c86044e21222ade366a32c019dc503d37165a621fe4a not found: ID does not exist" Oct 10 14:16:06 crc kubenswrapper[4745]: I1010 14:16:06.756937 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e33f867-88af-4331-b25f-1ac55b138e4d" path="/var/lib/kubelet/pods/7e33f867-88af-4331-b25f-1ac55b138e4d/volumes" Oct 10 14:16:16 crc kubenswrapper[4745]: I1010 14:16:16.186902 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:16:16 crc kubenswrapper[4745]: I1010 14:16:16.187533 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:16:30 crc kubenswrapper[4745]: I1010 14:16:30.639576 4745 generic.go:334] "Generic (PLEG): container finished" podID="5a7e2ddc-4584-43bf-8706-405a5f53b781" containerID="52dfac7e7ddc8facd62fb7de45ce63d37600f743df740eb3a4d7505170bdac35" exitCode=0 Oct 10 14:16:30 crc kubenswrapper[4745]: I1010 14:16:30.639648 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2mb5t/crc-debug-jcps2" event={"ID":"5a7e2ddc-4584-43bf-8706-405a5f53b781","Type":"ContainerDied","Data":"52dfac7e7ddc8facd62fb7de45ce63d37600f743df740eb3a4d7505170bdac35"} Oct 10 14:16:31 crc kubenswrapper[4745]: I1010 14:16:31.746544 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2mb5t/crc-debug-jcps2" Oct 10 14:16:31 crc kubenswrapper[4745]: I1010 14:16:31.776921 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2mb5t/crc-debug-jcps2"] Oct 10 14:16:31 crc kubenswrapper[4745]: I1010 14:16:31.788874 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2mb5t/crc-debug-jcps2"] Oct 10 14:16:31 crc kubenswrapper[4745]: I1010 14:16:31.900758 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t88tv\" (UniqueName: \"kubernetes.io/projected/5a7e2ddc-4584-43bf-8706-405a5f53b781-kube-api-access-t88tv\") pod \"5a7e2ddc-4584-43bf-8706-405a5f53b781\" (UID: \"5a7e2ddc-4584-43bf-8706-405a5f53b781\") " Oct 10 14:16:31 crc kubenswrapper[4745]: I1010 14:16:31.901168 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5a7e2ddc-4584-43bf-8706-405a5f53b781-host\") pod \"5a7e2ddc-4584-43bf-8706-405a5f53b781\" (UID: \"5a7e2ddc-4584-43bf-8706-405a5f53b781\") " Oct 10 14:16:31 crc kubenswrapper[4745]: I1010 14:16:31.901230 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a7e2ddc-4584-43bf-8706-405a5f53b781-host" (OuterVolumeSpecName: "host") pod "5a7e2ddc-4584-43bf-8706-405a5f53b781" (UID: "5a7e2ddc-4584-43bf-8706-405a5f53b781"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:16:31 crc kubenswrapper[4745]: I1010 14:16:31.901642 4745 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5a7e2ddc-4584-43bf-8706-405a5f53b781-host\") on node \"crc\" DevicePath \"\"" Oct 10 14:16:31 crc kubenswrapper[4745]: I1010 14:16:31.906045 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a7e2ddc-4584-43bf-8706-405a5f53b781-kube-api-access-t88tv" (OuterVolumeSpecName: "kube-api-access-t88tv") pod "5a7e2ddc-4584-43bf-8706-405a5f53b781" (UID: "5a7e2ddc-4584-43bf-8706-405a5f53b781"). InnerVolumeSpecName "kube-api-access-t88tv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:16:32 crc kubenswrapper[4745]: I1010 14:16:32.003155 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t88tv\" (UniqueName: \"kubernetes.io/projected/5a7e2ddc-4584-43bf-8706-405a5f53b781-kube-api-access-t88tv\") on node \"crc\" DevicePath \"\"" Oct 10 14:16:32 crc kubenswrapper[4745]: I1010 14:16:32.658748 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="189b9c77a5af4dd0721c0d650a1e96a94cbd8b1e046ea2bbc3ed282c6ff97f9c" Oct 10 14:16:32 crc kubenswrapper[4745]: I1010 14:16:32.659016 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2mb5t/crc-debug-jcps2" Oct 10 14:16:32 crc kubenswrapper[4745]: I1010 14:16:32.777533 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a7e2ddc-4584-43bf-8706-405a5f53b781" path="/var/lib/kubelet/pods/5a7e2ddc-4584-43bf-8706-405a5f53b781/volumes" Oct 10 14:16:32 crc kubenswrapper[4745]: I1010 14:16:32.970771 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2mb5t/crc-debug-vf6zp"] Oct 10 14:16:32 crc kubenswrapper[4745]: E1010 14:16:32.971338 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e33f867-88af-4331-b25f-1ac55b138e4d" containerName="extract-content" Oct 10 14:16:32 crc kubenswrapper[4745]: I1010 14:16:32.971415 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e33f867-88af-4331-b25f-1ac55b138e4d" containerName="extract-content" Oct 10 14:16:32 crc kubenswrapper[4745]: E1010 14:16:32.971489 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a7e2ddc-4584-43bf-8706-405a5f53b781" containerName="container-00" Oct 10 14:16:32 crc kubenswrapper[4745]: I1010 14:16:32.971549 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a7e2ddc-4584-43bf-8706-405a5f53b781" containerName="container-00" Oct 10 14:16:32 crc kubenswrapper[4745]: E1010 14:16:32.972168 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e33f867-88af-4331-b25f-1ac55b138e4d" containerName="extract-utilities" Oct 10 14:16:32 crc kubenswrapper[4745]: I1010 14:16:32.972236 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e33f867-88af-4331-b25f-1ac55b138e4d" containerName="extract-utilities" Oct 10 14:16:32 crc kubenswrapper[4745]: E1010 14:16:32.972319 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e33f867-88af-4331-b25f-1ac55b138e4d" containerName="registry-server" Oct 10 14:16:32 crc kubenswrapper[4745]: I1010 14:16:32.972379 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e33f867-88af-4331-b25f-1ac55b138e4d" containerName="registry-server" Oct 10 14:16:32 crc kubenswrapper[4745]: I1010 14:16:32.972616 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e33f867-88af-4331-b25f-1ac55b138e4d" containerName="registry-server" Oct 10 14:16:32 crc kubenswrapper[4745]: I1010 14:16:32.972706 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a7e2ddc-4584-43bf-8706-405a5f53b781" containerName="container-00" Oct 10 14:16:32 crc kubenswrapper[4745]: I1010 14:16:32.973384 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2mb5t/crc-debug-vf6zp" Oct 10 14:16:33 crc kubenswrapper[4745]: I1010 14:16:33.122582 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5sq6\" (UniqueName: \"kubernetes.io/projected/32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce-kube-api-access-t5sq6\") pod \"crc-debug-vf6zp\" (UID: \"32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce\") " pod="openshift-must-gather-2mb5t/crc-debug-vf6zp" Oct 10 14:16:33 crc kubenswrapper[4745]: I1010 14:16:33.122821 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce-host\") pod \"crc-debug-vf6zp\" (UID: \"32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce\") " pod="openshift-must-gather-2mb5t/crc-debug-vf6zp" Oct 10 14:16:33 crc kubenswrapper[4745]: I1010 14:16:33.224513 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce-host\") pod \"crc-debug-vf6zp\" (UID: \"32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce\") " pod="openshift-must-gather-2mb5t/crc-debug-vf6zp" Oct 10 14:16:33 crc kubenswrapper[4745]: I1010 14:16:33.224895 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5sq6\" (UniqueName: \"kubernetes.io/projected/32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce-kube-api-access-t5sq6\") pod \"crc-debug-vf6zp\" (UID: \"32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce\") " pod="openshift-must-gather-2mb5t/crc-debug-vf6zp" Oct 10 14:16:33 crc kubenswrapper[4745]: I1010 14:16:33.224705 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce-host\") pod \"crc-debug-vf6zp\" (UID: \"32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce\") " pod="openshift-must-gather-2mb5t/crc-debug-vf6zp" Oct 10 14:16:33 crc kubenswrapper[4745]: I1010 14:16:33.243457 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5sq6\" (UniqueName: \"kubernetes.io/projected/32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce-kube-api-access-t5sq6\") pod \"crc-debug-vf6zp\" (UID: \"32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce\") " pod="openshift-must-gather-2mb5t/crc-debug-vf6zp" Oct 10 14:16:33 crc kubenswrapper[4745]: I1010 14:16:33.288131 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2mb5t/crc-debug-vf6zp" Oct 10 14:16:33 crc kubenswrapper[4745]: I1010 14:16:33.667961 4745 generic.go:334] "Generic (PLEG): container finished" podID="32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce" containerID="83b76b5f1bf0d81e0268aa3d5e7b9a124cbcb9473dff3290c1569d5fcd5fbbc0" exitCode=0 Oct 10 14:16:33 crc kubenswrapper[4745]: I1010 14:16:33.668034 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2mb5t/crc-debug-vf6zp" event={"ID":"32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce","Type":"ContainerDied","Data":"83b76b5f1bf0d81e0268aa3d5e7b9a124cbcb9473dff3290c1569d5fcd5fbbc0"} Oct 10 14:16:33 crc kubenswrapper[4745]: I1010 14:16:33.668264 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2mb5t/crc-debug-vf6zp" event={"ID":"32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce","Type":"ContainerStarted","Data":"3623eec1776191a4392684341e28a703f337df8f07ed66bb2cdb19611b8b65e6"} Oct 10 14:16:34 crc kubenswrapper[4745]: I1010 14:16:34.020447 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2mb5t/crc-debug-vf6zp"] Oct 10 14:16:34 crc kubenswrapper[4745]: I1010 14:16:34.028532 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2mb5t/crc-debug-vf6zp"] Oct 10 14:16:34 crc kubenswrapper[4745]: I1010 14:16:34.779091 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2mb5t/crc-debug-vf6zp" Oct 10 14:16:34 crc kubenswrapper[4745]: I1010 14:16:34.956309 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce-host\") pod \"32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce\" (UID: \"32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce\") " Oct 10 14:16:34 crc kubenswrapper[4745]: I1010 14:16:34.956448 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce-host" (OuterVolumeSpecName: "host") pod "32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce" (UID: "32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:16:34 crc kubenswrapper[4745]: I1010 14:16:34.956480 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5sq6\" (UniqueName: \"kubernetes.io/projected/32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce-kube-api-access-t5sq6\") pod \"32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce\" (UID: \"32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce\") " Oct 10 14:16:34 crc kubenswrapper[4745]: I1010 14:16:34.956908 4745 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce-host\") on node \"crc\" DevicePath \"\"" Oct 10 14:16:34 crc kubenswrapper[4745]: I1010 14:16:34.963099 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce-kube-api-access-t5sq6" (OuterVolumeSpecName: "kube-api-access-t5sq6") pod "32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce" (UID: "32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce"). InnerVolumeSpecName "kube-api-access-t5sq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:16:35 crc kubenswrapper[4745]: I1010 14:16:35.059349 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5sq6\" (UniqueName: \"kubernetes.io/projected/32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce-kube-api-access-t5sq6\") on node \"crc\" DevicePath \"\"" Oct 10 14:16:35 crc kubenswrapper[4745]: I1010 14:16:35.186495 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2mb5t/crc-debug-fhf2k"] Oct 10 14:16:35 crc kubenswrapper[4745]: E1010 14:16:35.187062 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce" containerName="container-00" Oct 10 14:16:35 crc kubenswrapper[4745]: I1010 14:16:35.187096 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce" containerName="container-00" Oct 10 14:16:35 crc kubenswrapper[4745]: I1010 14:16:35.187363 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce" containerName="container-00" Oct 10 14:16:35 crc kubenswrapper[4745]: I1010 14:16:35.188126 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2mb5t/crc-debug-fhf2k" Oct 10 14:16:35 crc kubenswrapper[4745]: I1010 14:16:35.364605 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f4cdb212-86ab-4682-8712-c70cce45f426-host\") pod \"crc-debug-fhf2k\" (UID: \"f4cdb212-86ab-4682-8712-c70cce45f426\") " pod="openshift-must-gather-2mb5t/crc-debug-fhf2k" Oct 10 14:16:35 crc kubenswrapper[4745]: I1010 14:16:35.364852 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb6gj\" (UniqueName: \"kubernetes.io/projected/f4cdb212-86ab-4682-8712-c70cce45f426-kube-api-access-wb6gj\") pod \"crc-debug-fhf2k\" (UID: \"f4cdb212-86ab-4682-8712-c70cce45f426\") " pod="openshift-must-gather-2mb5t/crc-debug-fhf2k" Oct 10 14:16:35 crc kubenswrapper[4745]: I1010 14:16:35.466935 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb6gj\" (UniqueName: \"kubernetes.io/projected/f4cdb212-86ab-4682-8712-c70cce45f426-kube-api-access-wb6gj\") pod \"crc-debug-fhf2k\" (UID: \"f4cdb212-86ab-4682-8712-c70cce45f426\") " pod="openshift-must-gather-2mb5t/crc-debug-fhf2k" Oct 10 14:16:35 crc kubenswrapper[4745]: I1010 14:16:35.467024 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f4cdb212-86ab-4682-8712-c70cce45f426-host\") pod \"crc-debug-fhf2k\" (UID: \"f4cdb212-86ab-4682-8712-c70cce45f426\") " pod="openshift-must-gather-2mb5t/crc-debug-fhf2k" Oct 10 14:16:35 crc kubenswrapper[4745]: I1010 14:16:35.467199 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f4cdb212-86ab-4682-8712-c70cce45f426-host\") pod \"crc-debug-fhf2k\" (UID: \"f4cdb212-86ab-4682-8712-c70cce45f426\") " pod="openshift-must-gather-2mb5t/crc-debug-fhf2k" Oct 10 14:16:35 crc kubenswrapper[4745]: I1010 14:16:35.481962 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb6gj\" (UniqueName: \"kubernetes.io/projected/f4cdb212-86ab-4682-8712-c70cce45f426-kube-api-access-wb6gj\") pod \"crc-debug-fhf2k\" (UID: \"f4cdb212-86ab-4682-8712-c70cce45f426\") " pod="openshift-must-gather-2mb5t/crc-debug-fhf2k" Oct 10 14:16:35 crc kubenswrapper[4745]: I1010 14:16:35.508664 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2mb5t/crc-debug-fhf2k" Oct 10 14:16:35 crc kubenswrapper[4745]: I1010 14:16:35.688042 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2mb5t/crc-debug-vf6zp" Oct 10 14:16:35 crc kubenswrapper[4745]: I1010 14:16:35.688041 4745 scope.go:117] "RemoveContainer" containerID="83b76b5f1bf0d81e0268aa3d5e7b9a124cbcb9473dff3290c1569d5fcd5fbbc0" Oct 10 14:16:35 crc kubenswrapper[4745]: I1010 14:16:35.689200 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2mb5t/crc-debug-fhf2k" event={"ID":"f4cdb212-86ab-4682-8712-c70cce45f426","Type":"ContainerStarted","Data":"8478f1068f2d9786a4a81846ee604bffd771156d0c87d7216a28b1bea78984a6"} Oct 10 14:16:36 crc kubenswrapper[4745]: I1010 14:16:36.704133 4745 generic.go:334] "Generic (PLEG): container finished" podID="f4cdb212-86ab-4682-8712-c70cce45f426" containerID="e1137d9abefc625ad35fb8e19592cf3efd2aef854d7954080a4810dedf3e9616" exitCode=0 Oct 10 14:16:36 crc kubenswrapper[4745]: I1010 14:16:36.704192 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2mb5t/crc-debug-fhf2k" event={"ID":"f4cdb212-86ab-4682-8712-c70cce45f426","Type":"ContainerDied","Data":"e1137d9abefc625ad35fb8e19592cf3efd2aef854d7954080a4810dedf3e9616"} Oct 10 14:16:36 crc kubenswrapper[4745]: I1010 14:16:36.784258 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce" path="/var/lib/kubelet/pods/32fe3501-3cbd-4ba0-a2ba-d7f6cd7078ce/volumes" Oct 10 14:16:36 crc kubenswrapper[4745]: I1010 14:16:36.784953 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2mb5t/crc-debug-fhf2k"] Oct 10 14:16:36 crc kubenswrapper[4745]: I1010 14:16:36.784986 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2mb5t/crc-debug-fhf2k"] Oct 10 14:16:37 crc kubenswrapper[4745]: I1010 14:16:37.807327 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2mb5t/crc-debug-fhf2k" Oct 10 14:16:37 crc kubenswrapper[4745]: I1010 14:16:37.920492 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb6gj\" (UniqueName: \"kubernetes.io/projected/f4cdb212-86ab-4682-8712-c70cce45f426-kube-api-access-wb6gj\") pod \"f4cdb212-86ab-4682-8712-c70cce45f426\" (UID: \"f4cdb212-86ab-4682-8712-c70cce45f426\") " Oct 10 14:16:37 crc kubenswrapper[4745]: I1010 14:16:37.920760 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f4cdb212-86ab-4682-8712-c70cce45f426-host\") pod \"f4cdb212-86ab-4682-8712-c70cce45f426\" (UID: \"f4cdb212-86ab-4682-8712-c70cce45f426\") " Oct 10 14:16:37 crc kubenswrapper[4745]: I1010 14:16:37.922582 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4cdb212-86ab-4682-8712-c70cce45f426-host" (OuterVolumeSpecName: "host") pod "f4cdb212-86ab-4682-8712-c70cce45f426" (UID: "f4cdb212-86ab-4682-8712-c70cce45f426"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:16:37 crc kubenswrapper[4745]: I1010 14:16:37.938781 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4cdb212-86ab-4682-8712-c70cce45f426-kube-api-access-wb6gj" (OuterVolumeSpecName: "kube-api-access-wb6gj") pod "f4cdb212-86ab-4682-8712-c70cce45f426" (UID: "f4cdb212-86ab-4682-8712-c70cce45f426"). InnerVolumeSpecName "kube-api-access-wb6gj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:16:38 crc kubenswrapper[4745]: I1010 14:16:38.023805 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb6gj\" (UniqueName: \"kubernetes.io/projected/f4cdb212-86ab-4682-8712-c70cce45f426-kube-api-access-wb6gj\") on node \"crc\" DevicePath \"\"" Oct 10 14:16:38 crc kubenswrapper[4745]: I1010 14:16:38.023836 4745 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f4cdb212-86ab-4682-8712-c70cce45f426-host\") on node \"crc\" DevicePath \"\"" Oct 10 14:16:38 crc kubenswrapper[4745]: I1010 14:16:38.724524 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8478f1068f2d9786a4a81846ee604bffd771156d0c87d7216a28b1bea78984a6" Oct 10 14:16:38 crc kubenswrapper[4745]: I1010 14:16:38.724607 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2mb5t/crc-debug-fhf2k" Oct 10 14:16:38 crc kubenswrapper[4745]: I1010 14:16:38.766044 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4cdb212-86ab-4682-8712-c70cce45f426" path="/var/lib/kubelet/pods/f4cdb212-86ab-4682-8712-c70cce45f426/volumes" Oct 10 14:16:40 crc kubenswrapper[4745]: I1010 14:16:40.209547 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-795bd49d76-9gqws_5b21c656-1b66-422f-8397-12c630ee4889/barbican-api/0.log" Oct 10 14:16:40 crc kubenswrapper[4745]: I1010 14:16:40.368513 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-795bd49d76-9gqws_5b21c656-1b66-422f-8397-12c630ee4889/barbican-api-log/0.log" Oct 10 14:16:40 crc kubenswrapper[4745]: I1010 14:16:40.382291 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-568f47fdfd-mnxcr_2195e7f7-9227-404a-8123-3d487c93ac6b/barbican-keystone-listener/0.log" Oct 10 14:16:40 crc kubenswrapper[4745]: I1010 14:16:40.441097 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-568f47fdfd-mnxcr_2195e7f7-9227-404a-8123-3d487c93ac6b/barbican-keystone-listener-log/0.log" Oct 10 14:16:40 crc kubenswrapper[4745]: I1010 14:16:40.562580 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-57477cc59-r674m_82ee4ff2-4fa9-42fc-91dd-ee171c443ffa/barbican-worker/0.log" Oct 10 14:16:40 crc kubenswrapper[4745]: I1010 14:16:40.666822 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-57477cc59-r674m_82ee4ff2-4fa9-42fc-91dd-ee171c443ffa/barbican-worker-log/0.log" Oct 10 14:16:40 crc kubenswrapper[4745]: I1010 14:16:40.835020 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh_7f174876-f891-4508-97e6-e31624990bc8/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:16:40 crc kubenswrapper[4745]: I1010 14:16:40.888254 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0d791e6b-724b-4631-8a50-98b6dfd108f2/ceilometer-central-agent/0.log" Oct 10 14:16:40 crc kubenswrapper[4745]: I1010 14:16:40.928396 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0d791e6b-724b-4631-8a50-98b6dfd108f2/ceilometer-notification-agent/0.log" Oct 10 14:16:41 crc kubenswrapper[4745]: I1010 14:16:41.009235 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0d791e6b-724b-4631-8a50-98b6dfd108f2/proxy-httpd/0.log" Oct 10 14:16:41 crc kubenswrapper[4745]: I1010 14:16:41.016709 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0d791e6b-724b-4631-8a50-98b6dfd108f2/sg-core/0.log" Oct 10 14:16:41 crc kubenswrapper[4745]: I1010 14:16:41.150462 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_90b61af2-8fcd-40b6-8dc5-92207d219f45/cinder-api/0.log" Oct 10 14:16:41 crc kubenswrapper[4745]: I1010 14:16:41.205439 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_90b61af2-8fcd-40b6-8dc5-92207d219f45/cinder-api-log/0.log" Oct 10 14:16:41 crc kubenswrapper[4745]: I1010 14:16:41.380872 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_513c43ae-c96a-4ef0-8162-708490db52d6/probe/0.log" Oct 10 14:16:41 crc kubenswrapper[4745]: I1010 14:16:41.434909 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_513c43ae-c96a-4ef0-8162-708490db52d6/cinder-scheduler/0.log" Oct 10 14:16:41 crc kubenswrapper[4745]: I1010 14:16:41.532819 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-xprnx_986809df-787f-46ec-b4ac-dc28eaf82cc7/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:16:41 crc kubenswrapper[4745]: I1010 14:16:41.807143 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx_0a993d65-ad98-4529-95ad-1663fb206c5a/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:16:41 crc kubenswrapper[4745]: I1010 14:16:41.869988 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m_4329eb3b-cd56-417e-9da2-361ab8817091/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:16:41 crc kubenswrapper[4745]: I1010 14:16:41.986688 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-4hmj7_b5d53599-f2c8-4120-b767-72793898c000/init/0.log" Oct 10 14:16:42 crc kubenswrapper[4745]: I1010 14:16:42.154459 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-4hmj7_b5d53599-f2c8-4120-b767-72793898c000/init/0.log" Oct 10 14:16:42 crc kubenswrapper[4745]: I1010 14:16:42.228113 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-4hmj7_b5d53599-f2c8-4120-b767-72793898c000/dnsmasq-dns/0.log" Oct 10 14:16:42 crc kubenswrapper[4745]: I1010 14:16:42.233393 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8_e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:16:42 crc kubenswrapper[4745]: I1010 14:16:42.424837 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d2ba148a-0fdd-45d3-91e7-106c98cc7d2e/glance-httpd/0.log" Oct 10 14:16:42 crc kubenswrapper[4745]: I1010 14:16:42.460749 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d2ba148a-0fdd-45d3-91e7-106c98cc7d2e/glance-log/0.log" Oct 10 14:16:42 crc kubenswrapper[4745]: I1010 14:16:42.639450 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0a7607af-7365-45dc-9a8e-91719c20cdcf/glance-httpd/0.log" Oct 10 14:16:42 crc kubenswrapper[4745]: I1010 14:16:42.648749 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0a7607af-7365-45dc-9a8e-91719c20cdcf/glance-log/0.log" Oct 10 14:16:42 crc kubenswrapper[4745]: I1010 14:16:42.834552 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-8574589b48-pc9wr_fb40990e-1db6-4eab-9365-06678441a9ba/horizon/0.log" Oct 10 14:16:42 crc kubenswrapper[4745]: I1010 14:16:42.979413 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-rngcx_8360e97b-9df5-4790-8796-06c6280f323d/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:16:43 crc kubenswrapper[4745]: I1010 14:16:43.094600 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-8574589b48-pc9wr_fb40990e-1db6-4eab-9365-06678441a9ba/horizon-log/0.log" Oct 10 14:16:43 crc kubenswrapper[4745]: I1010 14:16:43.143945 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-hflsg_292290b6-c1e2-4b2b-955b-699aa826627d/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:16:43 crc kubenswrapper[4745]: I1010 14:16:43.407338 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29335081-sq6xb_a5f9c2a5-82c5-456e-b023-667c13b4d328/keystone-cron/0.log" Oct 10 14:16:43 crc kubenswrapper[4745]: I1010 14:16:43.462295 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-745bfbd6-ms6d2_a870bece-0e73-47f8-b0d2-79ee3e3cf637/keystone-api/0.log" Oct 10 14:16:43 crc kubenswrapper[4745]: I1010 14:16:43.627866 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_2c1adab6-47ba-4e16-b607-e61f92db1b53/kube-state-metrics/0.log" Oct 10 14:16:43 crc kubenswrapper[4745]: I1010 14:16:43.739302 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg_4709df73-fa0c-4b05-aa37-8d32a4aa52f1/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:16:44 crc kubenswrapper[4745]: I1010 14:16:44.011223 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-86b9f849df-zktvk_668e04d2-c914-4325-acab-c8638172e4cc/neutron-api/0.log" Oct 10 14:16:44 crc kubenswrapper[4745]: I1010 14:16:44.032699 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-86b9f849df-zktvk_668e04d2-c914-4325-acab-c8638172e4cc/neutron-httpd/0.log" Oct 10 14:16:44 crc kubenswrapper[4745]: I1010 14:16:44.145700 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2_4674795d-7ada-40eb-93d6-524c7af53ee3/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:16:44 crc kubenswrapper[4745]: I1010 14:16:44.747977 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_b1801b5b-c008-463f-952b-92ddef21faae/nova-api-log/0.log" Oct 10 14:16:44 crc kubenswrapper[4745]: I1010 14:16:44.761121 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_ac183214-b5c6-4e33-8045-ddef84c43323/nova-cell0-conductor-conductor/0.log" Oct 10 14:16:45 crc kubenswrapper[4745]: I1010 14:16:45.006301 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_b1801b5b-c008-463f-952b-92ddef21faae/nova-api-api/0.log" Oct 10 14:16:45 crc kubenswrapper[4745]: I1010 14:16:45.052750 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_87ac2f43-7e80-4310-8d3f-5750ae54cc06/nova-cell1-conductor-conductor/0.log" Oct 10 14:16:45 crc kubenswrapper[4745]: I1010 14:16:45.075238 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_6ff9dc11-4059-440b-a21f-8dd6ef566d36/nova-cell1-novncproxy-novncproxy/0.log" Oct 10 14:16:45 crc kubenswrapper[4745]: I1010 14:16:45.323897 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-nww5s_274ab6cc-4796-493d-ad35-5cf55195151c/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:16:45 crc kubenswrapper[4745]: I1010 14:16:45.456336 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_254ea8d2-b80e-40c0-8e33-3459e73878de/nova-metadata-log/0.log" Oct 10 14:16:45 crc kubenswrapper[4745]: I1010 14:16:45.726398 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_8d13d787-633b-4ae3-8fbd-013c3de88709/nova-scheduler-scheduler/0.log" Oct 10 14:16:45 crc kubenswrapper[4745]: I1010 14:16:45.784721 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c22070ab-5d89-42fa-b04a-640ec3045982/mysql-bootstrap/0.log" Oct 10 14:16:45 crc kubenswrapper[4745]: I1010 14:16:45.942399 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c22070ab-5d89-42fa-b04a-640ec3045982/mysql-bootstrap/0.log" Oct 10 14:16:45 crc kubenswrapper[4745]: I1010 14:16:45.989543 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c22070ab-5d89-42fa-b04a-640ec3045982/galera/0.log" Oct 10 14:16:46 crc kubenswrapper[4745]: I1010 14:16:46.179343 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_35e75435-406e-4053-a1c2-2a7fc992b31e/mysql-bootstrap/0.log" Oct 10 14:16:46 crc kubenswrapper[4745]: I1010 14:16:46.187315 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:16:46 crc kubenswrapper[4745]: I1010 14:16:46.187369 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:16:46 crc kubenswrapper[4745]: I1010 14:16:46.187427 4745 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 14:16:46 crc kubenswrapper[4745]: I1010 14:16:46.188138 4745 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264"} pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 14:16:46 crc kubenswrapper[4745]: I1010 14:16:46.188201 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" containerID="cri-o://a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" gracePeriod=600 Oct 10 14:16:46 crc kubenswrapper[4745]: E1010 14:16:46.311039 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:16:46 crc kubenswrapper[4745]: I1010 14:16:46.352297 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_35e75435-406e-4053-a1c2-2a7fc992b31e/mysql-bootstrap/0.log" Oct 10 14:16:46 crc kubenswrapper[4745]: I1010 14:16:46.385930 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_35e75435-406e-4053-a1c2-2a7fc992b31e/galera/0.log" Oct 10 14:16:46 crc kubenswrapper[4745]: I1010 14:16:46.556231 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_5cc91ec2-6a94-4923-82cf-d488c077a97b/openstackclient/0.log" Oct 10 14:16:46 crc kubenswrapper[4745]: I1010 14:16:46.665474 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-92pgn_513dcce5-f68e-42a4-ba94-44013577e96e/ovn-controller/0.log" Oct 10 14:16:46 crc kubenswrapper[4745]: I1010 14:16:46.711368 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_254ea8d2-b80e-40c0-8e33-3459e73878de/nova-metadata-metadata/0.log" Oct 10 14:16:46 crc kubenswrapper[4745]: I1010 14:16:46.816367 4745 generic.go:334] "Generic (PLEG): container finished" podID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" exitCode=0 Oct 10 14:16:46 crc kubenswrapper[4745]: I1010 14:16:46.816435 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerDied","Data":"a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264"} Oct 10 14:16:46 crc kubenswrapper[4745]: I1010 14:16:46.816475 4745 scope.go:117] "RemoveContainer" containerID="c465e4549e850aec30db19773f9543087be6ef4e34f1ddf2ef454efa7ce84172" Oct 10 14:16:46 crc kubenswrapper[4745]: I1010 14:16:46.817123 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:16:46 crc kubenswrapper[4745]: E1010 14:16:46.817462 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:16:46 crc kubenswrapper[4745]: I1010 14:16:46.846296 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-xhz46_3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce/openstack-network-exporter/0.log" Oct 10 14:16:46 crc kubenswrapper[4745]: I1010 14:16:46.979883 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s2qdr_e9b37d55-6160-43d6-8957-4cc31c95251d/ovsdb-server-init/0.log" Oct 10 14:16:47 crc kubenswrapper[4745]: I1010 14:16:47.139452 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s2qdr_e9b37d55-6160-43d6-8957-4cc31c95251d/ovs-vswitchd/0.log" Oct 10 14:16:47 crc kubenswrapper[4745]: I1010 14:16:47.149591 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s2qdr_e9b37d55-6160-43d6-8957-4cc31c95251d/ovsdb-server/0.log" Oct 10 14:16:47 crc kubenswrapper[4745]: I1010 14:16:47.157359 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s2qdr_e9b37d55-6160-43d6-8957-4cc31c95251d/ovsdb-server-init/0.log" Oct 10 14:16:47 crc kubenswrapper[4745]: I1010 14:16:47.364231 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_6f12c742-f26e-473c-ab59-02217e1b24c6/openstack-network-exporter/0.log" Oct 10 14:16:47 crc kubenswrapper[4745]: I1010 14:16:47.400836 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-5fz6h_ca7c907d-5166-4e04-b485-04d0b903d6cd/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:16:47 crc kubenswrapper[4745]: I1010 14:16:47.487969 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_6f12c742-f26e-473c-ab59-02217e1b24c6/ovn-northd/0.log" Oct 10 14:16:47 crc kubenswrapper[4745]: I1010 14:16:47.569144 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_dd428c33-2518-41ce-8dd4-b62fe3059525/openstack-network-exporter/0.log" Oct 10 14:16:47 crc kubenswrapper[4745]: I1010 14:16:47.609115 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_dd428c33-2518-41ce-8dd4-b62fe3059525/ovsdbserver-nb/0.log" Oct 10 14:16:47 crc kubenswrapper[4745]: I1010 14:16:47.769552 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_3894eba1-5402-4b7f-9859-a57827e29f2b/openstack-network-exporter/0.log" Oct 10 14:16:47 crc kubenswrapper[4745]: I1010 14:16:47.829482 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_3894eba1-5402-4b7f-9859-a57827e29f2b/ovsdbserver-sb/0.log" Oct 10 14:16:48 crc kubenswrapper[4745]: I1010 14:16:48.004049 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7b8bc58bc6-p44zs_5cf84964-c89c-4990-8320-823b10461186/placement-api/0.log" Oct 10 14:16:48 crc kubenswrapper[4745]: I1010 14:16:48.067685 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7b8bc58bc6-p44zs_5cf84964-c89c-4990-8320-823b10461186/placement-log/0.log" Oct 10 14:16:48 crc kubenswrapper[4745]: I1010 14:16:48.073963 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_972f7a23-bb59-4b6c-8ee7-a51647bbc261/setup-container/0.log" Oct 10 14:16:48 crc kubenswrapper[4745]: I1010 14:16:48.374908 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_972f7a23-bb59-4b6c-8ee7-a51647bbc261/rabbitmq/0.log" Oct 10 14:16:48 crc kubenswrapper[4745]: I1010 14:16:48.436336 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_972f7a23-bb59-4b6c-8ee7-a51647bbc261/setup-container/0.log" Oct 10 14:16:48 crc kubenswrapper[4745]: I1010 14:16:48.498969 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3f7d0aba-ccad-4d20-9ab1-cf7d080571c1/setup-container/0.log" Oct 10 14:16:48 crc kubenswrapper[4745]: I1010 14:16:48.605379 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3f7d0aba-ccad-4d20-9ab1-cf7d080571c1/setup-container/0.log" Oct 10 14:16:48 crc kubenswrapper[4745]: I1010 14:16:48.634173 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3f7d0aba-ccad-4d20-9ab1-cf7d080571c1/rabbitmq/0.log" Oct 10 14:16:48 crc kubenswrapper[4745]: I1010 14:16:48.726938 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n_b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:16:48 crc kubenswrapper[4745]: I1010 14:16:48.858924 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-fvppw_4fda6419-bf48-4f0e-8ff8-da6b9c2da25f/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:16:48 crc kubenswrapper[4745]: I1010 14:16:48.999838 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9_167c6043-6db1-4ead-8cf4-4040842b8d88/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:16:49 crc kubenswrapper[4745]: I1010 14:16:49.112264 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-vr2vp_29cef4fe-f039-411b-a979-91362f56cfd7/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:16:49 crc kubenswrapper[4745]: I1010 14:16:49.208296 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-2mslp_0d120415-3123-4b2b-a29d-b1b9c822ed7c/ssh-known-hosts-edpm-deployment/0.log" Oct 10 14:16:49 crc kubenswrapper[4745]: I1010 14:16:49.423143 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5d77644dd7-jv6wz_6aa7ade8-3d4b-4168-80a9-673c73b919a3/proxy-httpd/0.log" Oct 10 14:16:49 crc kubenswrapper[4745]: I1010 14:16:49.461336 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5d77644dd7-jv6wz_6aa7ade8-3d4b-4168-80a9-673c73b919a3/proxy-server/0.log" Oct 10 14:16:49 crc kubenswrapper[4745]: I1010 14:16:49.623180 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-8kzzn_749bd815-8704-47d6-b5a0-4e129ab976c1/swift-ring-rebalance/0.log" Oct 10 14:16:49 crc kubenswrapper[4745]: I1010 14:16:49.646668 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/account-auditor/0.log" Oct 10 14:16:49 crc kubenswrapper[4745]: I1010 14:16:49.797291 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/account-reaper/0.log" Oct 10 14:16:49 crc kubenswrapper[4745]: I1010 14:16:49.878374 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/account-replicator/0.log" Oct 10 14:16:49 crc kubenswrapper[4745]: I1010 14:16:49.911941 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/account-server/0.log" Oct 10 14:16:49 crc kubenswrapper[4745]: I1010 14:16:49.994293 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/container-auditor/0.log" Oct 10 14:16:50 crc kubenswrapper[4745]: I1010 14:16:50.079985 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/container-replicator/0.log" Oct 10 14:16:50 crc kubenswrapper[4745]: I1010 14:16:50.094995 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/container-server/0.log" Oct 10 14:16:50 crc kubenswrapper[4745]: I1010 14:16:50.138671 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/container-updater/0.log" Oct 10 14:16:50 crc kubenswrapper[4745]: I1010 14:16:50.206324 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/object-auditor/0.log" Oct 10 14:16:50 crc kubenswrapper[4745]: I1010 14:16:50.289618 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/object-expirer/0.log" Oct 10 14:16:50 crc kubenswrapper[4745]: I1010 14:16:50.360460 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/object-replicator/0.log" Oct 10 14:16:50 crc kubenswrapper[4745]: I1010 14:16:50.393906 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/object-server/0.log" Oct 10 14:16:50 crc kubenswrapper[4745]: I1010 14:16:50.443249 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/object-updater/0.log" Oct 10 14:16:50 crc kubenswrapper[4745]: I1010 14:16:50.474626 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/rsync/0.log" Oct 10 14:16:50 crc kubenswrapper[4745]: I1010 14:16:50.564640 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/swift-recon-cron/0.log" Oct 10 14:16:50 crc kubenswrapper[4745]: I1010 14:16:50.708916 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7_9c27594e-14e8-4b37-860c-19d4c47605cc/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:16:50 crc kubenswrapper[4745]: I1010 14:16:50.807013 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_55dab8c7-31e2-4694-9f9e-a3272376cddd/tempest-tests-tempest-tests-runner/0.log" Oct 10 14:16:50 crc kubenswrapper[4745]: I1010 14:16:50.955769 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_cb7b00b8-301c-4b88-a189-2019e53366ec/test-operator-logs-container/0.log" Oct 10 14:16:51 crc kubenswrapper[4745]: I1010 14:16:51.021088 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-b9582_e4f15e34-6886-448d-9ed1-7a5cbaf0fda2/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:16:57 crc kubenswrapper[4745]: I1010 14:16:57.744672 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:16:57 crc kubenswrapper[4745]: E1010 14:16:57.745352 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:16:59 crc kubenswrapper[4745]: I1010 14:16:59.044261 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_9e79afba-7f24-4e67-9c37-973256aac339/memcached/0.log" Oct 10 14:17:09 crc kubenswrapper[4745]: I1010 14:17:09.744843 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:17:09 crc kubenswrapper[4745]: E1010 14:17:09.745586 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:17:14 crc kubenswrapper[4745]: I1010 14:17:14.091172 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc_128069b5-6f89-4da5-9e35-bfe14f1ebd50/util/0.log" Oct 10 14:17:14 crc kubenswrapper[4745]: I1010 14:17:14.247633 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc_128069b5-6f89-4da5-9e35-bfe14f1ebd50/pull/0.log" Oct 10 14:17:14 crc kubenswrapper[4745]: I1010 14:17:14.264037 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc_128069b5-6f89-4da5-9e35-bfe14f1ebd50/pull/0.log" Oct 10 14:17:14 crc kubenswrapper[4745]: I1010 14:17:14.285413 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc_128069b5-6f89-4da5-9e35-bfe14f1ebd50/util/0.log" Oct 10 14:17:14 crc kubenswrapper[4745]: I1010 14:17:14.463389 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc_128069b5-6f89-4da5-9e35-bfe14f1ebd50/pull/0.log" Oct 10 14:17:14 crc kubenswrapper[4745]: I1010 14:17:14.473082 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc_128069b5-6f89-4da5-9e35-bfe14f1ebd50/util/0.log" Oct 10 14:17:14 crc kubenswrapper[4745]: I1010 14:17:14.487483 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc_128069b5-6f89-4da5-9e35-bfe14f1ebd50/extract/0.log" Oct 10 14:17:14 crc kubenswrapper[4745]: I1010 14:17:14.628904 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-htmcv_730b5555-5f6d-4e16-91e0-faacad56d777/kube-rbac-proxy/0.log" Oct 10 14:17:14 crc kubenswrapper[4745]: I1010 14:17:14.681094 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-c7986_5ca519af-249e-48f0-8669-77bd548be8f3/kube-rbac-proxy/0.log" Oct 10 14:17:14 crc kubenswrapper[4745]: I1010 14:17:14.759215 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-htmcv_730b5555-5f6d-4e16-91e0-faacad56d777/manager/0.log" Oct 10 14:17:14 crc kubenswrapper[4745]: I1010 14:17:14.842986 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-c7986_5ca519af-249e-48f0-8669-77bd548be8f3/manager/0.log" Oct 10 14:17:14 crc kubenswrapper[4745]: I1010 14:17:14.916960 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-fpcxk_dff847b8-915b-4b64-85c3-d7c20d1282a0/kube-rbac-proxy/0.log" Oct 10 14:17:14 crc kubenswrapper[4745]: I1010 14:17:14.960899 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-fpcxk_dff847b8-915b-4b64-85c3-d7c20d1282a0/manager/0.log" Oct 10 14:17:15 crc kubenswrapper[4745]: I1010 14:17:15.098867 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-tz665_446e0b63-81e5-47fb-b110-cfa5393e5311/kube-rbac-proxy/0.log" Oct 10 14:17:15 crc kubenswrapper[4745]: I1010 14:17:15.185067 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-tz665_446e0b63-81e5-47fb-b110-cfa5393e5311/manager/0.log" Oct 10 14:17:15 crc kubenswrapper[4745]: I1010 14:17:15.278786 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-2w6nw_7f07678c-0cc2-47a5-b244-72ae3aa55baa/kube-rbac-proxy/0.log" Oct 10 14:17:15 crc kubenswrapper[4745]: I1010 14:17:15.327893 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-2w6nw_7f07678c-0cc2-47a5-b244-72ae3aa55baa/manager/0.log" Oct 10 14:17:15 crc kubenswrapper[4745]: I1010 14:17:15.384935 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-qmbvh_1d2b124e-5495-42bd-b675-4eed70ccc446/kube-rbac-proxy/0.log" Oct 10 14:17:15 crc kubenswrapper[4745]: I1010 14:17:15.469184 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-qmbvh_1d2b124e-5495-42bd-b675-4eed70ccc446/manager/0.log" Oct 10 14:17:15 crc kubenswrapper[4745]: I1010 14:17:15.574259 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-dt9kx_443416f7-e4ce-41f2-b546-9344f0cb72a7/kube-rbac-proxy/0.log" Oct 10 14:17:15 crc kubenswrapper[4745]: I1010 14:17:15.724023 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-d5qnn_9fdf7ed4-1e18-43c0-b305-2b9b226f4634/kube-rbac-proxy/0.log" Oct 10 14:17:15 crc kubenswrapper[4745]: I1010 14:17:15.753589 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-dt9kx_443416f7-e4ce-41f2-b546-9344f0cb72a7/manager/0.log" Oct 10 14:17:15 crc kubenswrapper[4745]: I1010 14:17:15.772880 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-d5qnn_9fdf7ed4-1e18-43c0-b305-2b9b226f4634/manager/0.log" Oct 10 14:17:15 crc kubenswrapper[4745]: I1010 14:17:15.982690 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-h2bt5_10978ce4-0034-4ad4-9a47-2f1931c09655/kube-rbac-proxy/0.log" Oct 10 14:17:16 crc kubenswrapper[4745]: I1010 14:17:16.002340 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-h2bt5_10978ce4-0034-4ad4-9a47-2f1931c09655/manager/0.log" Oct 10 14:17:16 crc kubenswrapper[4745]: I1010 14:17:16.080024 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-t9wqw_fc401fc1-e03b-4b32-8782-15e9c2a70e8a/kube-rbac-proxy/0.log" Oct 10 14:17:16 crc kubenswrapper[4745]: I1010 14:17:16.173343 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-t9wqw_fc401fc1-e03b-4b32-8782-15e9c2a70e8a/manager/0.log" Oct 10 14:17:16 crc kubenswrapper[4745]: I1010 14:17:16.206822 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-vc9rk_420fdd11-0131-479d-9f7a-47f091cacd83/kube-rbac-proxy/0.log" Oct 10 14:17:16 crc kubenswrapper[4745]: I1010 14:17:16.270850 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-vc9rk_420fdd11-0131-479d-9f7a-47f091cacd83/manager/0.log" Oct 10 14:17:16 crc kubenswrapper[4745]: I1010 14:17:16.415200 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-2dmpc_83b0fe55-090f-448e-95c2-1d3cbbc1b54c/kube-rbac-proxy/0.log" Oct 10 14:17:16 crc kubenswrapper[4745]: I1010 14:17:16.451675 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-2dmpc_83b0fe55-090f-448e-95c2-1d3cbbc1b54c/manager/0.log" Oct 10 14:17:16 crc kubenswrapper[4745]: I1010 14:17:16.520982 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-89lzv_93defbbc-c732-4743-bcff-a1d7f06aefa2/kube-rbac-proxy/0.log" Oct 10 14:17:16 crc kubenswrapper[4745]: I1010 14:17:16.688839 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-89lzv_93defbbc-c732-4743-bcff-a1d7f06aefa2/manager/0.log" Oct 10 14:17:16 crc kubenswrapper[4745]: I1010 14:17:16.698597 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-xxvtr_e1c2fd77-699a-4c79-9c54-9251c5ad7d17/kube-rbac-proxy/0.log" Oct 10 14:17:16 crc kubenswrapper[4745]: I1010 14:17:16.802016 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-xxvtr_e1c2fd77-699a-4c79-9c54-9251c5ad7d17/manager/0.log" Oct 10 14:17:16 crc kubenswrapper[4745]: I1010 14:17:16.892685 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw_0174cdbc-06e0-423f-af63-5467737e6ec5/kube-rbac-proxy/0.log" Oct 10 14:17:16 crc kubenswrapper[4745]: I1010 14:17:16.953803 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw_0174cdbc-06e0-423f-af63-5467737e6ec5/manager/0.log" Oct 10 14:17:17 crc kubenswrapper[4745]: I1010 14:17:17.144002 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6f95c9794b-jtlvc_fd4f5298-2862-4650-950d-abd7f8195942/kube-rbac-proxy/0.log" Oct 10 14:17:17 crc kubenswrapper[4745]: I1010 14:17:17.227672 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7f65679b66-d82sn_b0e08ed6-f210-41b7-907e-d6722025f43e/kube-rbac-proxy/0.log" Oct 10 14:17:17 crc kubenswrapper[4745]: I1010 14:17:17.557581 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7f65679b66-d82sn_b0e08ed6-f210-41b7-907e-d6722025f43e/operator/0.log" Oct 10 14:17:17 crc kubenswrapper[4745]: I1010 14:17:17.611932 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-6tszb_40d5edac-be93-424b-8f39-1609f8b7cef4/registry-server/0.log" Oct 10 14:17:17 crc kubenswrapper[4745]: I1010 14:17:17.778163 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-rz48f_0ec7257b-b487-453c-88e8-d279184acdfc/kube-rbac-proxy/0.log" Oct 10 14:17:17 crc kubenswrapper[4745]: I1010 14:17:17.841927 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-rz48f_0ec7257b-b487-453c-88e8-d279184acdfc/manager/0.log" Oct 10 14:17:17 crc kubenswrapper[4745]: I1010 14:17:17.993795 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-b7w5k_36db8827-b9af-4b9b-87aa-3cfa7d005ee0/kube-rbac-proxy/0.log" Oct 10 14:17:18 crc kubenswrapper[4745]: I1010 14:17:18.004241 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-b7w5k_36db8827-b9af-4b9b-87aa-3cfa7d005ee0/manager/0.log" Oct 10 14:17:18 crc kubenswrapper[4745]: I1010 14:17:18.053064 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-c9gmp_ebc60df2-d5b0-418f-a8cd-e8440706d3be/operator/0.log" Oct 10 14:17:18 crc kubenswrapper[4745]: I1010 14:17:18.178450 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6f95c9794b-jtlvc_fd4f5298-2862-4650-950d-abd7f8195942/manager/0.log" Oct 10 14:17:18 crc kubenswrapper[4745]: I1010 14:17:18.197241 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-qzh4v_e6828db3-0a2c-4b98-8cbf-1eba04f13bc5/kube-rbac-proxy/0.log" Oct 10 14:17:18 crc kubenswrapper[4745]: I1010 14:17:18.333818 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-qzh4v_e6828db3-0a2c-4b98-8cbf-1eba04f13bc5/manager/0.log" Oct 10 14:17:18 crc kubenswrapper[4745]: I1010 14:17:18.401510 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-h4jn2_a5094dca-dde7-4144-bfbc-b4f90c170a21/kube-rbac-proxy/0.log" Oct 10 14:17:18 crc kubenswrapper[4745]: I1010 14:17:18.449100 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-h4jn2_a5094dca-dde7-4144-bfbc-b4f90c170a21/manager/0.log" Oct 10 14:17:18 crc kubenswrapper[4745]: I1010 14:17:18.478639 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-lr987_ba76d6db-10cc-4d7d-b620-e37533730395/kube-rbac-proxy/0.log" Oct 10 14:17:18 crc kubenswrapper[4745]: I1010 14:17:18.531388 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-lr987_ba76d6db-10cc-4d7d-b620-e37533730395/manager/0.log" Oct 10 14:17:18 crc kubenswrapper[4745]: I1010 14:17:18.607491 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-96mn7_6d02fe8c-f0b4-4ef1-8e57-569a836708cd/kube-rbac-proxy/0.log" Oct 10 14:17:18 crc kubenswrapper[4745]: I1010 14:17:18.705650 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-96mn7_6d02fe8c-f0b4-4ef1-8e57-569a836708cd/manager/0.log" Oct 10 14:17:20 crc kubenswrapper[4745]: I1010 14:17:20.745427 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:17:20 crc kubenswrapper[4745]: E1010 14:17:20.746185 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:17:32 crc kubenswrapper[4745]: I1010 14:17:32.745163 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:17:32 crc kubenswrapper[4745]: E1010 14:17:32.745916 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:17:32 crc kubenswrapper[4745]: I1010 14:17:32.832897 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-fqb4s_839da6a0-320a-45d1-8a6c-be916ee27421/control-plane-machine-set-operator/0.log" Oct 10 14:17:32 crc kubenswrapper[4745]: I1010 14:17:32.956283 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-s8wbg_800c6118-3c93-4dd4-b7e4-17f23b82de1d/machine-api-operator/0.log" Oct 10 14:17:32 crc kubenswrapper[4745]: I1010 14:17:32.976935 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-s8wbg_800c6118-3c93-4dd4-b7e4-17f23b82de1d/kube-rbac-proxy/0.log" Oct 10 14:17:44 crc kubenswrapper[4745]: I1010 14:17:44.291705 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-c24f6_61fb7cf8-4f9b-4414-824f-1f5270529840/cert-manager-controller/0.log" Oct 10 14:17:44 crc kubenswrapper[4745]: I1010 14:17:44.497307 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-rp25w_5f2820cb-2b97-4f6f-a92d-b6f16cb83de9/cert-manager-cainjector/0.log" Oct 10 14:17:44 crc kubenswrapper[4745]: I1010 14:17:44.519576 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-r4mqq_d365e19d-2015-4d4f-935e-79a4b3998dc4/cert-manager-webhook/0.log" Oct 10 14:17:47 crc kubenswrapper[4745]: I1010 14:17:47.744717 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:17:47 crc kubenswrapper[4745]: E1010 14:17:47.745623 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:17:55 crc kubenswrapper[4745]: I1010 14:17:55.590055 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-ws6j9_b8628d70-5b97-4c1c-b8c1-10e28a0439dd/nmstate-console-plugin/0.log" Oct 10 14:17:55 crc kubenswrapper[4745]: I1010 14:17:55.738161 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-6mcw2_9c18fb6e-a0ef-4367-a917-1abde904faf2/nmstate-handler/0.log" Oct 10 14:17:55 crc kubenswrapper[4745]: I1010 14:17:55.772053 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-bsftc_90c1fae7-ac4c-4d71-ac19-417d3f61f010/kube-rbac-proxy/0.log" Oct 10 14:17:55 crc kubenswrapper[4745]: I1010 14:17:55.790867 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-bsftc_90c1fae7-ac4c-4d71-ac19-417d3f61f010/nmstate-metrics/0.log" Oct 10 14:17:55 crc kubenswrapper[4745]: I1010 14:17:55.936154 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-8sw7s_f5837847-c246-4661-9174-b0ec21039982/nmstate-operator/0.log" Oct 10 14:17:55 crc kubenswrapper[4745]: I1010 14:17:55.973448 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-s6hqt_869f5cad-555c-4695-a9bc-f9e699484561/nmstate-webhook/0.log" Oct 10 14:18:00 crc kubenswrapper[4745]: I1010 14:18:00.745683 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:18:00 crc kubenswrapper[4745]: E1010 14:18:00.746468 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:18:09 crc kubenswrapper[4745]: I1010 14:18:09.245425 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-ktgdl_c1711da0-4710-4569-bdbd-2afe523b91be/kube-rbac-proxy/0.log" Oct 10 14:18:09 crc kubenswrapper[4745]: I1010 14:18:09.344935 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-ktgdl_c1711da0-4710-4569-bdbd-2afe523b91be/controller/0.log" Oct 10 14:18:09 crc kubenswrapper[4745]: I1010 14:18:09.484683 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-frr-files/0.log" Oct 10 14:18:09 crc kubenswrapper[4745]: I1010 14:18:09.708379 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-metrics/0.log" Oct 10 14:18:09 crc kubenswrapper[4745]: I1010 14:18:09.719121 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-frr-files/0.log" Oct 10 14:18:09 crc kubenswrapper[4745]: I1010 14:18:09.746053 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-reloader/0.log" Oct 10 14:18:09 crc kubenswrapper[4745]: I1010 14:18:09.751397 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-reloader/0.log" Oct 10 14:18:09 crc kubenswrapper[4745]: I1010 14:18:09.896894 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-reloader/0.log" Oct 10 14:18:09 crc kubenswrapper[4745]: I1010 14:18:09.901059 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-frr-files/0.log" Oct 10 14:18:09 crc kubenswrapper[4745]: I1010 14:18:09.929907 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-metrics/0.log" Oct 10 14:18:09 crc kubenswrapper[4745]: I1010 14:18:09.951299 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-metrics/0.log" Oct 10 14:18:10 crc kubenswrapper[4745]: I1010 14:18:10.126534 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-reloader/0.log" Oct 10 14:18:10 crc kubenswrapper[4745]: I1010 14:18:10.144283 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/controller/0.log" Oct 10 14:18:10 crc kubenswrapper[4745]: I1010 14:18:10.150069 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-metrics/0.log" Oct 10 14:18:10 crc kubenswrapper[4745]: I1010 14:18:10.150115 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-frr-files/0.log" Oct 10 14:18:10 crc kubenswrapper[4745]: I1010 14:18:10.334329 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/frr-metrics/0.log" Oct 10 14:18:10 crc kubenswrapper[4745]: I1010 14:18:10.340433 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/kube-rbac-proxy/0.log" Oct 10 14:18:10 crc kubenswrapper[4745]: I1010 14:18:10.390209 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/kube-rbac-proxy-frr/0.log" Oct 10 14:18:10 crc kubenswrapper[4745]: I1010 14:18:10.713411 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/reloader/0.log" Oct 10 14:18:10 crc kubenswrapper[4745]: I1010 14:18:10.848355 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-8ktnf_a362b7b7-7534-4a4f-ae57-aac84fd292c7/frr-k8s-webhook-server/0.log" Oct 10 14:18:10 crc kubenswrapper[4745]: I1010 14:18:10.975692 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6d67b47897-wgz9q_bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1/manager/0.log" Oct 10 14:18:11 crc kubenswrapper[4745]: I1010 14:18:11.300018 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-55796c598f-6xdbw_fcb7295e-30a8-4023-9f78-c75a0dadea04/webhook-server/0.log" Oct 10 14:18:11 crc kubenswrapper[4745]: I1010 14:18:11.336601 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-92z9w_b298d9ff-9b10-42f4-aabf-259673933a1a/kube-rbac-proxy/0.log" Oct 10 14:18:11 crc kubenswrapper[4745]: I1010 14:18:11.438944 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/frr/0.log" Oct 10 14:18:11 crc kubenswrapper[4745]: I1010 14:18:11.741352 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-92z9w_b298d9ff-9b10-42f4-aabf-259673933a1a/speaker/0.log" Oct 10 14:18:11 crc kubenswrapper[4745]: I1010 14:18:11.745506 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:18:11 crc kubenswrapper[4745]: E1010 14:18:11.745774 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:18:22 crc kubenswrapper[4745]: I1010 14:18:22.745147 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:18:22 crc kubenswrapper[4745]: E1010 14:18:22.745865 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:18:23 crc kubenswrapper[4745]: I1010 14:18:23.435927 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl_2050d745-31cf-44f8-8fb8-e1a4701c9799/util/0.log" Oct 10 14:18:23 crc kubenswrapper[4745]: I1010 14:18:23.650927 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl_2050d745-31cf-44f8-8fb8-e1a4701c9799/util/0.log" Oct 10 14:18:23 crc kubenswrapper[4745]: I1010 14:18:23.659816 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl_2050d745-31cf-44f8-8fb8-e1a4701c9799/pull/0.log" Oct 10 14:18:23 crc kubenswrapper[4745]: I1010 14:18:23.662601 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl_2050d745-31cf-44f8-8fb8-e1a4701c9799/pull/0.log" Oct 10 14:18:23 crc kubenswrapper[4745]: I1010 14:18:23.862886 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl_2050d745-31cf-44f8-8fb8-e1a4701c9799/extract/0.log" Oct 10 14:18:23 crc kubenswrapper[4745]: I1010 14:18:23.894429 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl_2050d745-31cf-44f8-8fb8-e1a4701c9799/util/0.log" Oct 10 14:18:23 crc kubenswrapper[4745]: I1010 14:18:23.913301 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl_2050d745-31cf-44f8-8fb8-e1a4701c9799/pull/0.log" Oct 10 14:18:24 crc kubenswrapper[4745]: I1010 14:18:24.017786 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h5nbz_78f0f7b2-023b-4227-a64a-c3ccae73349b/extract-utilities/0.log" Oct 10 14:18:24 crc kubenswrapper[4745]: I1010 14:18:24.196000 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h5nbz_78f0f7b2-023b-4227-a64a-c3ccae73349b/extract-content/0.log" Oct 10 14:18:24 crc kubenswrapper[4745]: I1010 14:18:24.196954 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h5nbz_78f0f7b2-023b-4227-a64a-c3ccae73349b/extract-content/0.log" Oct 10 14:18:24 crc kubenswrapper[4745]: I1010 14:18:24.240761 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h5nbz_78f0f7b2-023b-4227-a64a-c3ccae73349b/extract-utilities/0.log" Oct 10 14:18:24 crc kubenswrapper[4745]: I1010 14:18:24.402276 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h5nbz_78f0f7b2-023b-4227-a64a-c3ccae73349b/extract-content/0.log" Oct 10 14:18:24 crc kubenswrapper[4745]: I1010 14:18:24.403448 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h5nbz_78f0f7b2-023b-4227-a64a-c3ccae73349b/extract-utilities/0.log" Oct 10 14:18:24 crc kubenswrapper[4745]: I1010 14:18:24.614427 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-stdb8_ba2a7571-2a26-4bf9-a642-ef927ed9c7f8/extract-utilities/0.log" Oct 10 14:18:24 crc kubenswrapper[4745]: I1010 14:18:24.810544 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-stdb8_ba2a7571-2a26-4bf9-a642-ef927ed9c7f8/extract-content/0.log" Oct 10 14:18:24 crc kubenswrapper[4745]: I1010 14:18:24.839432 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-stdb8_ba2a7571-2a26-4bf9-a642-ef927ed9c7f8/extract-utilities/0.log" Oct 10 14:18:24 crc kubenswrapper[4745]: I1010 14:18:24.857026 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-stdb8_ba2a7571-2a26-4bf9-a642-ef927ed9c7f8/extract-content/0.log" Oct 10 14:18:24 crc kubenswrapper[4745]: I1010 14:18:24.921929 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h5nbz_78f0f7b2-023b-4227-a64a-c3ccae73349b/registry-server/0.log" Oct 10 14:18:25 crc kubenswrapper[4745]: I1010 14:18:25.030348 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-stdb8_ba2a7571-2a26-4bf9-a642-ef927ed9c7f8/extract-utilities/0.log" Oct 10 14:18:25 crc kubenswrapper[4745]: I1010 14:18:25.063184 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-stdb8_ba2a7571-2a26-4bf9-a642-ef927ed9c7f8/extract-content/0.log" Oct 10 14:18:25 crc kubenswrapper[4745]: I1010 14:18:25.266583 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c_4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4/util/0.log" Oct 10 14:18:25 crc kubenswrapper[4745]: I1010 14:18:25.406648 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c_4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4/pull/0.log" Oct 10 14:18:25 crc kubenswrapper[4745]: I1010 14:18:25.452323 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c_4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4/util/0.log" Oct 10 14:18:25 crc kubenswrapper[4745]: I1010 14:18:25.494116 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c_4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4/pull/0.log" Oct 10 14:18:25 crc kubenswrapper[4745]: I1010 14:18:25.499821 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-stdb8_ba2a7571-2a26-4bf9-a642-ef927ed9c7f8/registry-server/0.log" Oct 10 14:18:25 crc kubenswrapper[4745]: I1010 14:18:25.646043 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c_4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4/extract/0.log" Oct 10 14:18:25 crc kubenswrapper[4745]: I1010 14:18:25.666949 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c_4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4/util/0.log" Oct 10 14:18:25 crc kubenswrapper[4745]: I1010 14:18:25.667386 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c_4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4/pull/0.log" Oct 10 14:18:25 crc kubenswrapper[4745]: I1010 14:18:25.804299 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hbrh4_a87ec403-812a-4af1-83eb-95788143fcc4/marketplace-operator/0.log" Oct 10 14:18:25 crc kubenswrapper[4745]: I1010 14:18:25.874481 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bwpkj_511552f9-6c48-4742-9332-a5724f4d5697/extract-utilities/0.log" Oct 10 14:18:26 crc kubenswrapper[4745]: I1010 14:18:26.028295 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bwpkj_511552f9-6c48-4742-9332-a5724f4d5697/extract-content/0.log" Oct 10 14:18:26 crc kubenswrapper[4745]: I1010 14:18:26.041435 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bwpkj_511552f9-6c48-4742-9332-a5724f4d5697/extract-content/0.log" Oct 10 14:18:26 crc kubenswrapper[4745]: I1010 14:18:26.060232 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bwpkj_511552f9-6c48-4742-9332-a5724f4d5697/extract-utilities/0.log" Oct 10 14:18:26 crc kubenswrapper[4745]: I1010 14:18:26.217396 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bwpkj_511552f9-6c48-4742-9332-a5724f4d5697/extract-utilities/0.log" Oct 10 14:18:26 crc kubenswrapper[4745]: I1010 14:18:26.222217 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bwpkj_511552f9-6c48-4742-9332-a5724f4d5697/extract-content/0.log" Oct 10 14:18:26 crc kubenswrapper[4745]: I1010 14:18:26.341125 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bwpkj_511552f9-6c48-4742-9332-a5724f4d5697/registry-server/0.log" Oct 10 14:18:26 crc kubenswrapper[4745]: I1010 14:18:26.406271 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k7bwk_c3284f0e-0f23-4b38-bc27-3b2806155073/extract-utilities/0.log" Oct 10 14:18:26 crc kubenswrapper[4745]: I1010 14:18:26.558603 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k7bwk_c3284f0e-0f23-4b38-bc27-3b2806155073/extract-content/0.log" Oct 10 14:18:26 crc kubenswrapper[4745]: I1010 14:18:26.560153 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k7bwk_c3284f0e-0f23-4b38-bc27-3b2806155073/extract-utilities/0.log" Oct 10 14:18:26 crc kubenswrapper[4745]: I1010 14:18:26.595378 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k7bwk_c3284f0e-0f23-4b38-bc27-3b2806155073/extract-content/0.log" Oct 10 14:18:26 crc kubenswrapper[4745]: I1010 14:18:26.760364 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k7bwk_c3284f0e-0f23-4b38-bc27-3b2806155073/extract-content/0.log" Oct 10 14:18:26 crc kubenswrapper[4745]: I1010 14:18:26.769547 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k7bwk_c3284f0e-0f23-4b38-bc27-3b2806155073/extract-utilities/0.log" Oct 10 14:18:27 crc kubenswrapper[4745]: I1010 14:18:27.357006 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k7bwk_c3284f0e-0f23-4b38-bc27-3b2806155073/registry-server/0.log" Oct 10 14:18:36 crc kubenswrapper[4745]: I1010 14:18:36.750609 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:18:36 crc kubenswrapper[4745]: E1010 14:18:36.751319 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:18:50 crc kubenswrapper[4745]: E1010 14:18:50.589333 4745 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.200:53368->38.102.83.200:36957: write tcp 38.102.83.200:53368->38.102.83.200:36957: write: broken pipe Oct 10 14:18:51 crc kubenswrapper[4745]: I1010 14:18:51.748100 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:18:51 crc kubenswrapper[4745]: E1010 14:18:51.749033 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:18:58 crc kubenswrapper[4745]: E1010 14:18:58.532792 4745 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.200:42228->38.102.83.200:36957: write tcp 38.102.83.200:42228->38.102.83.200:36957: write: broken pipe Oct 10 14:19:02 crc kubenswrapper[4745]: I1010 14:19:02.745561 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:19:02 crc kubenswrapper[4745]: E1010 14:19:02.746192 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:19:14 crc kubenswrapper[4745]: I1010 14:19:14.744548 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:19:14 crc kubenswrapper[4745]: E1010 14:19:14.745351 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:19:29 crc kubenswrapper[4745]: I1010 14:19:29.744584 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:19:29 crc kubenswrapper[4745]: E1010 14:19:29.745631 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:19:40 crc kubenswrapper[4745]: I1010 14:19:40.745645 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:19:40 crc kubenswrapper[4745]: E1010 14:19:40.746267 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:19:53 crc kubenswrapper[4745]: I1010 14:19:53.745712 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:19:53 crc kubenswrapper[4745]: E1010 14:19:53.746872 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:20:02 crc kubenswrapper[4745]: I1010 14:20:02.679154 4745 generic.go:334] "Generic (PLEG): container finished" podID="59d60977-5951-405c-ac97-7c9d6dfd68f5" containerID="d2211773dcc78db2c4a4cc8dd94a0b34073237015e0021d0825c6cecf525ac9c" exitCode=0 Oct 10 14:20:02 crc kubenswrapper[4745]: I1010 14:20:02.679274 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2mb5t/must-gather-rkp6m" event={"ID":"59d60977-5951-405c-ac97-7c9d6dfd68f5","Type":"ContainerDied","Data":"d2211773dcc78db2c4a4cc8dd94a0b34073237015e0021d0825c6cecf525ac9c"} Oct 10 14:20:02 crc kubenswrapper[4745]: I1010 14:20:02.680841 4745 scope.go:117] "RemoveContainer" containerID="d2211773dcc78db2c4a4cc8dd94a0b34073237015e0021d0825c6cecf525ac9c" Oct 10 14:20:03 crc kubenswrapper[4745]: I1010 14:20:03.379370 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2mb5t_must-gather-rkp6m_59d60977-5951-405c-ac97-7c9d6dfd68f5/gather/0.log" Oct 10 14:20:06 crc kubenswrapper[4745]: I1010 14:20:06.761611 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:20:06 crc kubenswrapper[4745]: E1010 14:20:06.762616 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:20:10 crc kubenswrapper[4745]: I1010 14:20:10.904464 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2mb5t/must-gather-rkp6m"] Oct 10 14:20:10 crc kubenswrapper[4745]: I1010 14:20:10.905307 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-2mb5t/must-gather-rkp6m" podUID="59d60977-5951-405c-ac97-7c9d6dfd68f5" containerName="copy" containerID="cri-o://a2171f4b84612b69d263d5b3c0bce6c81aaa9378cb05a01b7a4a071174eb5183" gracePeriod=2 Oct 10 14:20:10 crc kubenswrapper[4745]: I1010 14:20:10.927589 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2mb5t/must-gather-rkp6m"] Oct 10 14:20:11 crc kubenswrapper[4745]: I1010 14:20:11.451340 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2mb5t_must-gather-rkp6m_59d60977-5951-405c-ac97-7c9d6dfd68f5/copy/0.log" Oct 10 14:20:11 crc kubenswrapper[4745]: I1010 14:20:11.452204 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2mb5t/must-gather-rkp6m" Oct 10 14:20:11 crc kubenswrapper[4745]: I1010 14:20:11.632832 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sql7p\" (UniqueName: \"kubernetes.io/projected/59d60977-5951-405c-ac97-7c9d6dfd68f5-kube-api-access-sql7p\") pod \"59d60977-5951-405c-ac97-7c9d6dfd68f5\" (UID: \"59d60977-5951-405c-ac97-7c9d6dfd68f5\") " Oct 10 14:20:11 crc kubenswrapper[4745]: I1010 14:20:11.632876 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/59d60977-5951-405c-ac97-7c9d6dfd68f5-must-gather-output\") pod \"59d60977-5951-405c-ac97-7c9d6dfd68f5\" (UID: \"59d60977-5951-405c-ac97-7c9d6dfd68f5\") " Oct 10 14:20:11 crc kubenswrapper[4745]: I1010 14:20:11.641559 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59d60977-5951-405c-ac97-7c9d6dfd68f5-kube-api-access-sql7p" (OuterVolumeSpecName: "kube-api-access-sql7p") pod "59d60977-5951-405c-ac97-7c9d6dfd68f5" (UID: "59d60977-5951-405c-ac97-7c9d6dfd68f5"). InnerVolumeSpecName "kube-api-access-sql7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:20:11 crc kubenswrapper[4745]: I1010 14:20:11.735777 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sql7p\" (UniqueName: \"kubernetes.io/projected/59d60977-5951-405c-ac97-7c9d6dfd68f5-kube-api-access-sql7p\") on node \"crc\" DevicePath \"\"" Oct 10 14:20:11 crc kubenswrapper[4745]: I1010 14:20:11.764271 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2mb5t_must-gather-rkp6m_59d60977-5951-405c-ac97-7c9d6dfd68f5/copy/0.log" Oct 10 14:20:11 crc kubenswrapper[4745]: I1010 14:20:11.765539 4745 generic.go:334] "Generic (PLEG): container finished" podID="59d60977-5951-405c-ac97-7c9d6dfd68f5" containerID="a2171f4b84612b69d263d5b3c0bce6c81aaa9378cb05a01b7a4a071174eb5183" exitCode=143 Oct 10 14:20:11 crc kubenswrapper[4745]: I1010 14:20:11.765598 4745 scope.go:117] "RemoveContainer" containerID="a2171f4b84612b69d263d5b3c0bce6c81aaa9378cb05a01b7a4a071174eb5183" Oct 10 14:20:11 crc kubenswrapper[4745]: I1010 14:20:11.765716 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2mb5t/must-gather-rkp6m" Oct 10 14:20:11 crc kubenswrapper[4745]: I1010 14:20:11.783826 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59d60977-5951-405c-ac97-7c9d6dfd68f5-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "59d60977-5951-405c-ac97-7c9d6dfd68f5" (UID: "59d60977-5951-405c-ac97-7c9d6dfd68f5"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:20:11 crc kubenswrapper[4745]: I1010 14:20:11.807548 4745 scope.go:117] "RemoveContainer" containerID="d2211773dcc78db2c4a4cc8dd94a0b34073237015e0021d0825c6cecf525ac9c" Oct 10 14:20:11 crc kubenswrapper[4745]: I1010 14:20:11.837591 4745 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/59d60977-5951-405c-ac97-7c9d6dfd68f5-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 10 14:20:11 crc kubenswrapper[4745]: I1010 14:20:11.898488 4745 scope.go:117] "RemoveContainer" containerID="a2171f4b84612b69d263d5b3c0bce6c81aaa9378cb05a01b7a4a071174eb5183" Oct 10 14:20:11 crc kubenswrapper[4745]: E1010 14:20:11.912963 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2171f4b84612b69d263d5b3c0bce6c81aaa9378cb05a01b7a4a071174eb5183\": container with ID starting with a2171f4b84612b69d263d5b3c0bce6c81aaa9378cb05a01b7a4a071174eb5183 not found: ID does not exist" containerID="a2171f4b84612b69d263d5b3c0bce6c81aaa9378cb05a01b7a4a071174eb5183" Oct 10 14:20:11 crc kubenswrapper[4745]: I1010 14:20:11.913008 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2171f4b84612b69d263d5b3c0bce6c81aaa9378cb05a01b7a4a071174eb5183"} err="failed to get container status \"a2171f4b84612b69d263d5b3c0bce6c81aaa9378cb05a01b7a4a071174eb5183\": rpc error: code = NotFound desc = could not find container \"a2171f4b84612b69d263d5b3c0bce6c81aaa9378cb05a01b7a4a071174eb5183\": container with ID starting with a2171f4b84612b69d263d5b3c0bce6c81aaa9378cb05a01b7a4a071174eb5183 not found: ID does not exist" Oct 10 14:20:11 crc kubenswrapper[4745]: I1010 14:20:11.913035 4745 scope.go:117] "RemoveContainer" containerID="d2211773dcc78db2c4a4cc8dd94a0b34073237015e0021d0825c6cecf525ac9c" Oct 10 14:20:11 crc kubenswrapper[4745]: E1010 14:20:11.916056 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2211773dcc78db2c4a4cc8dd94a0b34073237015e0021d0825c6cecf525ac9c\": container with ID starting with d2211773dcc78db2c4a4cc8dd94a0b34073237015e0021d0825c6cecf525ac9c not found: ID does not exist" containerID="d2211773dcc78db2c4a4cc8dd94a0b34073237015e0021d0825c6cecf525ac9c" Oct 10 14:20:11 crc kubenswrapper[4745]: I1010 14:20:11.916091 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2211773dcc78db2c4a4cc8dd94a0b34073237015e0021d0825c6cecf525ac9c"} err="failed to get container status \"d2211773dcc78db2c4a4cc8dd94a0b34073237015e0021d0825c6cecf525ac9c\": rpc error: code = NotFound desc = could not find container \"d2211773dcc78db2c4a4cc8dd94a0b34073237015e0021d0825c6cecf525ac9c\": container with ID starting with d2211773dcc78db2c4a4cc8dd94a0b34073237015e0021d0825c6cecf525ac9c not found: ID does not exist" Oct 10 14:20:12 crc kubenswrapper[4745]: I1010 14:20:12.755238 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59d60977-5951-405c-ac97-7c9d6dfd68f5" path="/var/lib/kubelet/pods/59d60977-5951-405c-ac97-7c9d6dfd68f5/volumes" Oct 10 14:20:19 crc kubenswrapper[4745]: I1010 14:20:19.745382 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:20:19 crc kubenswrapper[4745]: E1010 14:20:19.746360 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:20:30 crc kubenswrapper[4745]: I1010 14:20:30.745149 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:20:30 crc kubenswrapper[4745]: E1010 14:20:30.746088 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:20:41 crc kubenswrapper[4745]: I1010 14:20:41.745075 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:20:41 crc kubenswrapper[4745]: E1010 14:20:41.747710 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:20:50 crc kubenswrapper[4745]: I1010 14:20:50.835962 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mns4j/must-gather-6c92c"] Oct 10 14:20:50 crc kubenswrapper[4745]: E1010 14:20:50.836981 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4cdb212-86ab-4682-8712-c70cce45f426" containerName="container-00" Oct 10 14:20:50 crc kubenswrapper[4745]: I1010 14:20:50.836995 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4cdb212-86ab-4682-8712-c70cce45f426" containerName="container-00" Oct 10 14:20:50 crc kubenswrapper[4745]: E1010 14:20:50.837036 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59d60977-5951-405c-ac97-7c9d6dfd68f5" containerName="copy" Oct 10 14:20:50 crc kubenswrapper[4745]: I1010 14:20:50.837045 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="59d60977-5951-405c-ac97-7c9d6dfd68f5" containerName="copy" Oct 10 14:20:50 crc kubenswrapper[4745]: E1010 14:20:50.837057 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59d60977-5951-405c-ac97-7c9d6dfd68f5" containerName="gather" Oct 10 14:20:50 crc kubenswrapper[4745]: I1010 14:20:50.837064 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="59d60977-5951-405c-ac97-7c9d6dfd68f5" containerName="gather" Oct 10 14:20:50 crc kubenswrapper[4745]: I1010 14:20:50.837284 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="59d60977-5951-405c-ac97-7c9d6dfd68f5" containerName="copy" Oct 10 14:20:50 crc kubenswrapper[4745]: I1010 14:20:50.837293 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="59d60977-5951-405c-ac97-7c9d6dfd68f5" containerName="gather" Oct 10 14:20:50 crc kubenswrapper[4745]: I1010 14:20:50.837301 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4cdb212-86ab-4682-8712-c70cce45f426" containerName="container-00" Oct 10 14:20:50 crc kubenswrapper[4745]: I1010 14:20:50.838254 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mns4j/must-gather-6c92c" Oct 10 14:20:50 crc kubenswrapper[4745]: I1010 14:20:50.849594 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mns4j/must-gather-6c92c"] Oct 10 14:20:50 crc kubenswrapper[4745]: I1010 14:20:50.850874 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mns4j"/"kube-root-ca.crt" Oct 10 14:20:50 crc kubenswrapper[4745]: I1010 14:20:50.851539 4745 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mns4j"/"openshift-service-ca.crt" Oct 10 14:20:51 crc kubenswrapper[4745]: I1010 14:20:51.005647 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/19294049-0379-4bd4-bdaa-c46d50052501-must-gather-output\") pod \"must-gather-6c92c\" (UID: \"19294049-0379-4bd4-bdaa-c46d50052501\") " pod="openshift-must-gather-mns4j/must-gather-6c92c" Oct 10 14:20:51 crc kubenswrapper[4745]: I1010 14:20:51.006580 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvpjj\" (UniqueName: \"kubernetes.io/projected/19294049-0379-4bd4-bdaa-c46d50052501-kube-api-access-bvpjj\") pod \"must-gather-6c92c\" (UID: \"19294049-0379-4bd4-bdaa-c46d50052501\") " pod="openshift-must-gather-mns4j/must-gather-6c92c" Oct 10 14:20:51 crc kubenswrapper[4745]: I1010 14:20:51.108584 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvpjj\" (UniqueName: \"kubernetes.io/projected/19294049-0379-4bd4-bdaa-c46d50052501-kube-api-access-bvpjj\") pod \"must-gather-6c92c\" (UID: \"19294049-0379-4bd4-bdaa-c46d50052501\") " pod="openshift-must-gather-mns4j/must-gather-6c92c" Oct 10 14:20:51 crc kubenswrapper[4745]: I1010 14:20:51.108675 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/19294049-0379-4bd4-bdaa-c46d50052501-must-gather-output\") pod \"must-gather-6c92c\" (UID: \"19294049-0379-4bd4-bdaa-c46d50052501\") " pod="openshift-must-gather-mns4j/must-gather-6c92c" Oct 10 14:20:51 crc kubenswrapper[4745]: I1010 14:20:51.109206 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/19294049-0379-4bd4-bdaa-c46d50052501-must-gather-output\") pod \"must-gather-6c92c\" (UID: \"19294049-0379-4bd4-bdaa-c46d50052501\") " pod="openshift-must-gather-mns4j/must-gather-6c92c" Oct 10 14:20:51 crc kubenswrapper[4745]: I1010 14:20:51.137594 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvpjj\" (UniqueName: \"kubernetes.io/projected/19294049-0379-4bd4-bdaa-c46d50052501-kube-api-access-bvpjj\") pod \"must-gather-6c92c\" (UID: \"19294049-0379-4bd4-bdaa-c46d50052501\") " pod="openshift-must-gather-mns4j/must-gather-6c92c" Oct 10 14:20:51 crc kubenswrapper[4745]: I1010 14:20:51.159474 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mns4j/must-gather-6c92c" Oct 10 14:20:52 crc kubenswrapper[4745]: I1010 14:20:52.229457 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mns4j/must-gather-6c92c"] Oct 10 14:20:53 crc kubenswrapper[4745]: I1010 14:20:53.173709 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mns4j/must-gather-6c92c" event={"ID":"19294049-0379-4bd4-bdaa-c46d50052501","Type":"ContainerStarted","Data":"a69f0f0c0f239785f5fc334c8faa6c6d1ee6660f38b299d5917c040e1dfa9347"} Oct 10 14:20:53 crc kubenswrapper[4745]: I1010 14:20:53.174849 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mns4j/must-gather-6c92c" event={"ID":"19294049-0379-4bd4-bdaa-c46d50052501","Type":"ContainerStarted","Data":"3b94cccace5d99f025668855489c519fe976bed9dca92d6d811f8f458152e094"} Oct 10 14:20:53 crc kubenswrapper[4745]: I1010 14:20:53.174888 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mns4j/must-gather-6c92c" event={"ID":"19294049-0379-4bd4-bdaa-c46d50052501","Type":"ContainerStarted","Data":"6b13cae89d1c5a39d4a2702ff1e0ad7c3fd19821fb15fbc5730404804aa29e89"} Oct 10 14:20:53 crc kubenswrapper[4745]: I1010 14:20:53.196712 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mns4j/must-gather-6c92c" podStartSLOduration=3.196672847 podStartE2EDuration="3.196672847s" podCreationTimestamp="2025-10-10 14:20:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:20:53.196158244 +0000 UTC m=+3767.093815027" watchObservedRunningTime="2025-10-10 14:20:53.196672847 +0000 UTC m=+3767.094329620" Oct 10 14:20:53 crc kubenswrapper[4745]: I1010 14:20:53.744607 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:20:53 crc kubenswrapper[4745]: E1010 14:20:53.746048 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:20:56 crc kubenswrapper[4745]: I1010 14:20:56.034901 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mns4j/crc-debug-hxq4r"] Oct 10 14:20:56 crc kubenswrapper[4745]: I1010 14:20:56.036858 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mns4j/crc-debug-hxq4r" Oct 10 14:20:56 crc kubenswrapper[4745]: I1010 14:20:56.041708 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mns4j"/"default-dockercfg-g5mvb" Oct 10 14:20:56 crc kubenswrapper[4745]: I1010 14:20:56.110006 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e2784f2c-c690-4d54-af71-08887016a310-host\") pod \"crc-debug-hxq4r\" (UID: \"e2784f2c-c690-4d54-af71-08887016a310\") " pod="openshift-must-gather-mns4j/crc-debug-hxq4r" Oct 10 14:20:56 crc kubenswrapper[4745]: I1010 14:20:56.110068 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xclz\" (UniqueName: \"kubernetes.io/projected/e2784f2c-c690-4d54-af71-08887016a310-kube-api-access-4xclz\") pod \"crc-debug-hxq4r\" (UID: \"e2784f2c-c690-4d54-af71-08887016a310\") " pod="openshift-must-gather-mns4j/crc-debug-hxq4r" Oct 10 14:20:56 crc kubenswrapper[4745]: I1010 14:20:56.212780 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e2784f2c-c690-4d54-af71-08887016a310-host\") pod \"crc-debug-hxq4r\" (UID: \"e2784f2c-c690-4d54-af71-08887016a310\") " pod="openshift-must-gather-mns4j/crc-debug-hxq4r" Oct 10 14:20:56 crc kubenswrapper[4745]: I1010 14:20:56.212826 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xclz\" (UniqueName: \"kubernetes.io/projected/e2784f2c-c690-4d54-af71-08887016a310-kube-api-access-4xclz\") pod \"crc-debug-hxq4r\" (UID: \"e2784f2c-c690-4d54-af71-08887016a310\") " pod="openshift-must-gather-mns4j/crc-debug-hxq4r" Oct 10 14:20:56 crc kubenswrapper[4745]: I1010 14:20:56.213162 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e2784f2c-c690-4d54-af71-08887016a310-host\") pod \"crc-debug-hxq4r\" (UID: \"e2784f2c-c690-4d54-af71-08887016a310\") " pod="openshift-must-gather-mns4j/crc-debug-hxq4r" Oct 10 14:20:56 crc kubenswrapper[4745]: I1010 14:20:56.240324 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xclz\" (UniqueName: \"kubernetes.io/projected/e2784f2c-c690-4d54-af71-08887016a310-kube-api-access-4xclz\") pod \"crc-debug-hxq4r\" (UID: \"e2784f2c-c690-4d54-af71-08887016a310\") " pod="openshift-must-gather-mns4j/crc-debug-hxq4r" Oct 10 14:20:56 crc kubenswrapper[4745]: I1010 14:20:56.367037 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mns4j/crc-debug-hxq4r" Oct 10 14:20:56 crc kubenswrapper[4745]: W1010 14:20:56.402762 4745 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2784f2c_c690_4d54_af71_08887016a310.slice/crio-bcc6c655a25c9b7a65a3acf65a942345a4cd2fb19f1550a2ee363e3d7f1cb667 WatchSource:0}: Error finding container bcc6c655a25c9b7a65a3acf65a942345a4cd2fb19f1550a2ee363e3d7f1cb667: Status 404 returned error can't find the container with id bcc6c655a25c9b7a65a3acf65a942345a4cd2fb19f1550a2ee363e3d7f1cb667 Oct 10 14:20:57 crc kubenswrapper[4745]: I1010 14:20:57.264312 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mns4j/crc-debug-hxq4r" event={"ID":"e2784f2c-c690-4d54-af71-08887016a310","Type":"ContainerStarted","Data":"389a1b800d641a9b9abf749b5854bd3ce0e7f9617e7653944ac38cd5c77448f7"} Oct 10 14:20:57 crc kubenswrapper[4745]: I1010 14:20:57.264874 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mns4j/crc-debug-hxq4r" event={"ID":"e2784f2c-c690-4d54-af71-08887016a310","Type":"ContainerStarted","Data":"bcc6c655a25c9b7a65a3acf65a942345a4cd2fb19f1550a2ee363e3d7f1cb667"} Oct 10 14:20:57 crc kubenswrapper[4745]: I1010 14:20:57.280015 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mns4j/crc-debug-hxq4r" podStartSLOduration=1.2800006480000001 podStartE2EDuration="1.280000648s" podCreationTimestamp="2025-10-10 14:20:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:20:57.278248773 +0000 UTC m=+3771.175905536" watchObservedRunningTime="2025-10-10 14:20:57.280000648 +0000 UTC m=+3771.177657401" Oct 10 14:21:07 crc kubenswrapper[4745]: I1010 14:21:07.745367 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:21:07 crc kubenswrapper[4745]: E1010 14:21:07.746501 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:21:22 crc kubenswrapper[4745]: I1010 14:21:22.744809 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:21:22 crc kubenswrapper[4745]: E1010 14:21:22.745503 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:21:30 crc kubenswrapper[4745]: I1010 14:21:30.554821 4745 generic.go:334] "Generic (PLEG): container finished" podID="e2784f2c-c690-4d54-af71-08887016a310" containerID="389a1b800d641a9b9abf749b5854bd3ce0e7f9617e7653944ac38cd5c77448f7" exitCode=0 Oct 10 14:21:30 crc kubenswrapper[4745]: I1010 14:21:30.554900 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mns4j/crc-debug-hxq4r" event={"ID":"e2784f2c-c690-4d54-af71-08887016a310","Type":"ContainerDied","Data":"389a1b800d641a9b9abf749b5854bd3ce0e7f9617e7653944ac38cd5c77448f7"} Oct 10 14:21:31 crc kubenswrapper[4745]: I1010 14:21:31.667546 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mns4j/crc-debug-hxq4r" Oct 10 14:21:31 crc kubenswrapper[4745]: I1010 14:21:31.700456 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mns4j/crc-debug-hxq4r"] Oct 10 14:21:31 crc kubenswrapper[4745]: I1010 14:21:31.710275 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mns4j/crc-debug-hxq4r"] Oct 10 14:21:31 crc kubenswrapper[4745]: I1010 14:21:31.808233 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xclz\" (UniqueName: \"kubernetes.io/projected/e2784f2c-c690-4d54-af71-08887016a310-kube-api-access-4xclz\") pod \"e2784f2c-c690-4d54-af71-08887016a310\" (UID: \"e2784f2c-c690-4d54-af71-08887016a310\") " Oct 10 14:21:31 crc kubenswrapper[4745]: I1010 14:21:31.808650 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e2784f2c-c690-4d54-af71-08887016a310-host\") pod \"e2784f2c-c690-4d54-af71-08887016a310\" (UID: \"e2784f2c-c690-4d54-af71-08887016a310\") " Oct 10 14:21:31 crc kubenswrapper[4745]: I1010 14:21:31.809297 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2784f2c-c690-4d54-af71-08887016a310-host" (OuterVolumeSpecName: "host") pod "e2784f2c-c690-4d54-af71-08887016a310" (UID: "e2784f2c-c690-4d54-af71-08887016a310"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:21:31 crc kubenswrapper[4745]: I1010 14:21:31.809526 4745 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e2784f2c-c690-4d54-af71-08887016a310-host\") on node \"crc\" DevicePath \"\"" Oct 10 14:21:31 crc kubenswrapper[4745]: I1010 14:21:31.815145 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2784f2c-c690-4d54-af71-08887016a310-kube-api-access-4xclz" (OuterVolumeSpecName: "kube-api-access-4xclz") pod "e2784f2c-c690-4d54-af71-08887016a310" (UID: "e2784f2c-c690-4d54-af71-08887016a310"). InnerVolumeSpecName "kube-api-access-4xclz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:21:31 crc kubenswrapper[4745]: I1010 14:21:31.960818 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xclz\" (UniqueName: \"kubernetes.io/projected/e2784f2c-c690-4d54-af71-08887016a310-kube-api-access-4xclz\") on node \"crc\" DevicePath \"\"" Oct 10 14:21:32 crc kubenswrapper[4745]: I1010 14:21:32.571354 4745 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcc6c655a25c9b7a65a3acf65a942345a4cd2fb19f1550a2ee363e3d7f1cb667" Oct 10 14:21:32 crc kubenswrapper[4745]: I1010 14:21:32.571428 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mns4j/crc-debug-hxq4r" Oct 10 14:21:32 crc kubenswrapper[4745]: I1010 14:21:32.762046 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2784f2c-c690-4d54-af71-08887016a310" path="/var/lib/kubelet/pods/e2784f2c-c690-4d54-af71-08887016a310/volumes" Oct 10 14:21:32 crc kubenswrapper[4745]: I1010 14:21:32.873660 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mns4j/crc-debug-cmzp7"] Oct 10 14:21:32 crc kubenswrapper[4745]: E1010 14:21:32.874376 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2784f2c-c690-4d54-af71-08887016a310" containerName="container-00" Oct 10 14:21:32 crc kubenswrapper[4745]: I1010 14:21:32.874416 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2784f2c-c690-4d54-af71-08887016a310" containerName="container-00" Oct 10 14:21:32 crc kubenswrapper[4745]: I1010 14:21:32.874895 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2784f2c-c690-4d54-af71-08887016a310" containerName="container-00" Oct 10 14:21:32 crc kubenswrapper[4745]: I1010 14:21:32.876325 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mns4j/crc-debug-cmzp7" Oct 10 14:21:32 crc kubenswrapper[4745]: I1010 14:21:32.878911 4745 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mns4j"/"default-dockercfg-g5mvb" Oct 10 14:21:32 crc kubenswrapper[4745]: I1010 14:21:32.977329 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bbb5f4ce-6309-4652-81e8-e83fdff58046-host\") pod \"crc-debug-cmzp7\" (UID: \"bbb5f4ce-6309-4652-81e8-e83fdff58046\") " pod="openshift-must-gather-mns4j/crc-debug-cmzp7" Oct 10 14:21:32 crc kubenswrapper[4745]: I1010 14:21:32.977796 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n98mm\" (UniqueName: \"kubernetes.io/projected/bbb5f4ce-6309-4652-81e8-e83fdff58046-kube-api-access-n98mm\") pod \"crc-debug-cmzp7\" (UID: \"bbb5f4ce-6309-4652-81e8-e83fdff58046\") " pod="openshift-must-gather-mns4j/crc-debug-cmzp7" Oct 10 14:21:33 crc kubenswrapper[4745]: I1010 14:21:33.080103 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n98mm\" (UniqueName: \"kubernetes.io/projected/bbb5f4ce-6309-4652-81e8-e83fdff58046-kube-api-access-n98mm\") pod \"crc-debug-cmzp7\" (UID: \"bbb5f4ce-6309-4652-81e8-e83fdff58046\") " pod="openshift-must-gather-mns4j/crc-debug-cmzp7" Oct 10 14:21:33 crc kubenswrapper[4745]: I1010 14:21:33.080499 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bbb5f4ce-6309-4652-81e8-e83fdff58046-host\") pod \"crc-debug-cmzp7\" (UID: \"bbb5f4ce-6309-4652-81e8-e83fdff58046\") " pod="openshift-must-gather-mns4j/crc-debug-cmzp7" Oct 10 14:21:33 crc kubenswrapper[4745]: I1010 14:21:33.080688 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bbb5f4ce-6309-4652-81e8-e83fdff58046-host\") pod \"crc-debug-cmzp7\" (UID: \"bbb5f4ce-6309-4652-81e8-e83fdff58046\") " pod="openshift-must-gather-mns4j/crc-debug-cmzp7" Oct 10 14:21:33 crc kubenswrapper[4745]: I1010 14:21:33.095899 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n98mm\" (UniqueName: \"kubernetes.io/projected/bbb5f4ce-6309-4652-81e8-e83fdff58046-kube-api-access-n98mm\") pod \"crc-debug-cmzp7\" (UID: \"bbb5f4ce-6309-4652-81e8-e83fdff58046\") " pod="openshift-must-gather-mns4j/crc-debug-cmzp7" Oct 10 14:21:33 crc kubenswrapper[4745]: I1010 14:21:33.193425 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mns4j/crc-debug-cmzp7" Oct 10 14:21:33 crc kubenswrapper[4745]: I1010 14:21:33.581628 4745 generic.go:334] "Generic (PLEG): container finished" podID="bbb5f4ce-6309-4652-81e8-e83fdff58046" containerID="9d5e524b76a5a94603a3be868b05f7b2cc89a82484d11fdc28cf4becaf0e4f71" exitCode=0 Oct 10 14:21:33 crc kubenswrapper[4745]: I1010 14:21:33.581682 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mns4j/crc-debug-cmzp7" event={"ID":"bbb5f4ce-6309-4652-81e8-e83fdff58046","Type":"ContainerDied","Data":"9d5e524b76a5a94603a3be868b05f7b2cc89a82484d11fdc28cf4becaf0e4f71"} Oct 10 14:21:33 crc kubenswrapper[4745]: I1010 14:21:33.581722 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mns4j/crc-debug-cmzp7" event={"ID":"bbb5f4ce-6309-4652-81e8-e83fdff58046","Type":"ContainerStarted","Data":"d6c4df4781068318044b6d83119fca9d8d65f3d197cd5d148061991dbf5d469f"} Oct 10 14:21:34 crc kubenswrapper[4745]: I1010 14:21:34.014599 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mns4j/crc-debug-cmzp7"] Oct 10 14:21:34 crc kubenswrapper[4745]: I1010 14:21:34.027318 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mns4j/crc-debug-cmzp7"] Oct 10 14:21:34 crc kubenswrapper[4745]: I1010 14:21:34.693583 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mns4j/crc-debug-cmzp7" Oct 10 14:21:34 crc kubenswrapper[4745]: I1010 14:21:34.824844 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bbb5f4ce-6309-4652-81e8-e83fdff58046-host\") pod \"bbb5f4ce-6309-4652-81e8-e83fdff58046\" (UID: \"bbb5f4ce-6309-4652-81e8-e83fdff58046\") " Oct 10 14:21:34 crc kubenswrapper[4745]: I1010 14:21:34.825041 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n98mm\" (UniqueName: \"kubernetes.io/projected/bbb5f4ce-6309-4652-81e8-e83fdff58046-kube-api-access-n98mm\") pod \"bbb5f4ce-6309-4652-81e8-e83fdff58046\" (UID: \"bbb5f4ce-6309-4652-81e8-e83fdff58046\") " Oct 10 14:21:34 crc kubenswrapper[4745]: I1010 14:21:34.825210 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bbb5f4ce-6309-4652-81e8-e83fdff58046-host" (OuterVolumeSpecName: "host") pod "bbb5f4ce-6309-4652-81e8-e83fdff58046" (UID: "bbb5f4ce-6309-4652-81e8-e83fdff58046"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:21:34 crc kubenswrapper[4745]: I1010 14:21:34.825627 4745 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bbb5f4ce-6309-4652-81e8-e83fdff58046-host\") on node \"crc\" DevicePath \"\"" Oct 10 14:21:34 crc kubenswrapper[4745]: I1010 14:21:34.831166 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbb5f4ce-6309-4652-81e8-e83fdff58046-kube-api-access-n98mm" (OuterVolumeSpecName: "kube-api-access-n98mm") pod "bbb5f4ce-6309-4652-81e8-e83fdff58046" (UID: "bbb5f4ce-6309-4652-81e8-e83fdff58046"). InnerVolumeSpecName "kube-api-access-n98mm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:21:34 crc kubenswrapper[4745]: I1010 14:21:34.928260 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n98mm\" (UniqueName: \"kubernetes.io/projected/bbb5f4ce-6309-4652-81e8-e83fdff58046-kube-api-access-n98mm\") on node \"crc\" DevicePath \"\"" Oct 10 14:21:35 crc kubenswrapper[4745]: I1010 14:21:35.193275 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mns4j/crc-debug-xx29r"] Oct 10 14:21:35 crc kubenswrapper[4745]: E1010 14:21:35.193663 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbb5f4ce-6309-4652-81e8-e83fdff58046" containerName="container-00" Oct 10 14:21:35 crc kubenswrapper[4745]: I1010 14:21:35.193675 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbb5f4ce-6309-4652-81e8-e83fdff58046" containerName="container-00" Oct 10 14:21:35 crc kubenswrapper[4745]: I1010 14:21:35.193873 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbb5f4ce-6309-4652-81e8-e83fdff58046" containerName="container-00" Oct 10 14:21:35 crc kubenswrapper[4745]: I1010 14:21:35.194439 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mns4j/crc-debug-xx29r" Oct 10 14:21:35 crc kubenswrapper[4745]: I1010 14:21:35.235206 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b69b971f-6f8a-4486-abe7-9052b04587fd-host\") pod \"crc-debug-xx29r\" (UID: \"b69b971f-6f8a-4486-abe7-9052b04587fd\") " pod="openshift-must-gather-mns4j/crc-debug-xx29r" Oct 10 14:21:35 crc kubenswrapper[4745]: I1010 14:21:35.235446 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhlnh\" (UniqueName: \"kubernetes.io/projected/b69b971f-6f8a-4486-abe7-9052b04587fd-kube-api-access-bhlnh\") pod \"crc-debug-xx29r\" (UID: \"b69b971f-6f8a-4486-abe7-9052b04587fd\") " pod="openshift-must-gather-mns4j/crc-debug-xx29r" Oct 10 14:21:35 crc kubenswrapper[4745]: I1010 14:21:35.337175 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b69b971f-6f8a-4486-abe7-9052b04587fd-host\") pod \"crc-debug-xx29r\" (UID: \"b69b971f-6f8a-4486-abe7-9052b04587fd\") " pod="openshift-must-gather-mns4j/crc-debug-xx29r" Oct 10 14:21:35 crc kubenswrapper[4745]: I1010 14:21:35.337256 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhlnh\" (UniqueName: \"kubernetes.io/projected/b69b971f-6f8a-4486-abe7-9052b04587fd-kube-api-access-bhlnh\") pod \"crc-debug-xx29r\" (UID: \"b69b971f-6f8a-4486-abe7-9052b04587fd\") " pod="openshift-must-gather-mns4j/crc-debug-xx29r" Oct 10 14:21:35 crc kubenswrapper[4745]: I1010 14:21:35.337409 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b69b971f-6f8a-4486-abe7-9052b04587fd-host\") pod \"crc-debug-xx29r\" (UID: \"b69b971f-6f8a-4486-abe7-9052b04587fd\") " pod="openshift-must-gather-mns4j/crc-debug-xx29r" Oct 10 14:21:35 crc kubenswrapper[4745]: I1010 14:21:35.354891 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhlnh\" (UniqueName: \"kubernetes.io/projected/b69b971f-6f8a-4486-abe7-9052b04587fd-kube-api-access-bhlnh\") pod \"crc-debug-xx29r\" (UID: \"b69b971f-6f8a-4486-abe7-9052b04587fd\") " pod="openshift-must-gather-mns4j/crc-debug-xx29r" Oct 10 14:21:35 crc kubenswrapper[4745]: I1010 14:21:35.514390 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mns4j/crc-debug-xx29r" Oct 10 14:21:35 crc kubenswrapper[4745]: I1010 14:21:35.609752 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mns4j/crc-debug-xx29r" event={"ID":"b69b971f-6f8a-4486-abe7-9052b04587fd","Type":"ContainerStarted","Data":"1170847c1390347877a81edb688fda847ecf11c602e77baf7953551f410d69ff"} Oct 10 14:21:35 crc kubenswrapper[4745]: I1010 14:21:35.611363 4745 scope.go:117] "RemoveContainer" containerID="9d5e524b76a5a94603a3be868b05f7b2cc89a82484d11fdc28cf4becaf0e4f71" Oct 10 14:21:35 crc kubenswrapper[4745]: I1010 14:21:35.611610 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mns4j/crc-debug-cmzp7" Oct 10 14:21:35 crc kubenswrapper[4745]: E1010 14:21:35.749219 4745 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbb5f4ce_6309_4652_81e8_e83fdff58046.slice/crio-d6c4df4781068318044b6d83119fca9d8d65f3d197cd5d148061991dbf5d469f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbb5f4ce_6309_4652_81e8_e83fdff58046.slice\": RecentStats: unable to find data in memory cache]" Oct 10 14:21:36 crc kubenswrapper[4745]: I1010 14:21:36.622349 4745 generic.go:334] "Generic (PLEG): container finished" podID="b69b971f-6f8a-4486-abe7-9052b04587fd" containerID="9d8b349f15c842709fef448ba7f94b5370b3aeb4f1d13acfa1eeb34fe3a61adc" exitCode=0 Oct 10 14:21:36 crc kubenswrapper[4745]: I1010 14:21:36.622469 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mns4j/crc-debug-xx29r" event={"ID":"b69b971f-6f8a-4486-abe7-9052b04587fd","Type":"ContainerDied","Data":"9d8b349f15c842709fef448ba7f94b5370b3aeb4f1d13acfa1eeb34fe3a61adc"} Oct 10 14:21:36 crc kubenswrapper[4745]: I1010 14:21:36.654590 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mns4j/crc-debug-xx29r"] Oct 10 14:21:36 crc kubenswrapper[4745]: I1010 14:21:36.666790 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mns4j/crc-debug-xx29r"] Oct 10 14:21:36 crc kubenswrapper[4745]: I1010 14:21:36.755484 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbb5f4ce-6309-4652-81e8-e83fdff58046" path="/var/lib/kubelet/pods/bbb5f4ce-6309-4652-81e8-e83fdff58046/volumes" Oct 10 14:21:37 crc kubenswrapper[4745]: I1010 14:21:37.727934 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mns4j/crc-debug-xx29r" Oct 10 14:21:37 crc kubenswrapper[4745]: I1010 14:21:37.745627 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:21:37 crc kubenswrapper[4745]: E1010 14:21:37.746063 4745 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dnq2_openshift-machine-config-operator(5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" Oct 10 14:21:37 crc kubenswrapper[4745]: I1010 14:21:37.905868 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhlnh\" (UniqueName: \"kubernetes.io/projected/b69b971f-6f8a-4486-abe7-9052b04587fd-kube-api-access-bhlnh\") pod \"b69b971f-6f8a-4486-abe7-9052b04587fd\" (UID: \"b69b971f-6f8a-4486-abe7-9052b04587fd\") " Oct 10 14:21:37 crc kubenswrapper[4745]: I1010 14:21:37.906208 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b69b971f-6f8a-4486-abe7-9052b04587fd-host\") pod \"b69b971f-6f8a-4486-abe7-9052b04587fd\" (UID: \"b69b971f-6f8a-4486-abe7-9052b04587fd\") " Oct 10 14:21:37 crc kubenswrapper[4745]: I1010 14:21:37.906333 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b69b971f-6f8a-4486-abe7-9052b04587fd-host" (OuterVolumeSpecName: "host") pod "b69b971f-6f8a-4486-abe7-9052b04587fd" (UID: "b69b971f-6f8a-4486-abe7-9052b04587fd"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:21:37 crc kubenswrapper[4745]: I1010 14:21:37.906668 4745 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b69b971f-6f8a-4486-abe7-9052b04587fd-host\") on node \"crc\" DevicePath \"\"" Oct 10 14:21:37 crc kubenswrapper[4745]: I1010 14:21:37.912436 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b69b971f-6f8a-4486-abe7-9052b04587fd-kube-api-access-bhlnh" (OuterVolumeSpecName: "kube-api-access-bhlnh") pod "b69b971f-6f8a-4486-abe7-9052b04587fd" (UID: "b69b971f-6f8a-4486-abe7-9052b04587fd"). InnerVolumeSpecName "kube-api-access-bhlnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:21:38 crc kubenswrapper[4745]: I1010 14:21:38.007544 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhlnh\" (UniqueName: \"kubernetes.io/projected/b69b971f-6f8a-4486-abe7-9052b04587fd-kube-api-access-bhlnh\") on node \"crc\" DevicePath \"\"" Oct 10 14:21:38 crc kubenswrapper[4745]: I1010 14:21:38.642235 4745 scope.go:117] "RemoveContainer" containerID="9d8b349f15c842709fef448ba7f94b5370b3aeb4f1d13acfa1eeb34fe3a61adc" Oct 10 14:21:38 crc kubenswrapper[4745]: I1010 14:21:38.642305 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mns4j/crc-debug-xx29r" Oct 10 14:21:38 crc kubenswrapper[4745]: I1010 14:21:38.755720 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b69b971f-6f8a-4486-abe7-9052b04587fd" path="/var/lib/kubelet/pods/b69b971f-6f8a-4486-abe7-9052b04587fd/volumes" Oct 10 14:21:48 crc kubenswrapper[4745]: I1010 14:21:48.745696 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:21:49 crc kubenswrapper[4745]: I1010 14:21:49.630130 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-795bd49d76-9gqws_5b21c656-1b66-422f-8397-12c630ee4889/barbican-api/0.log" Oct 10 14:21:49 crc kubenswrapper[4745]: I1010 14:21:49.761110 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"4e4dc030f0d245a23b97a0415e5a9289fcfa1518fa939913ff0ab29c429c1984"} Oct 10 14:21:49 crc kubenswrapper[4745]: I1010 14:21:49.765501 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-795bd49d76-9gqws_5b21c656-1b66-422f-8397-12c630ee4889/barbican-api-log/0.log" Oct 10 14:21:49 crc kubenswrapper[4745]: I1010 14:21:49.821411 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-568f47fdfd-mnxcr_2195e7f7-9227-404a-8123-3d487c93ac6b/barbican-keystone-listener/0.log" Oct 10 14:21:49 crc kubenswrapper[4745]: I1010 14:21:49.889011 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-568f47fdfd-mnxcr_2195e7f7-9227-404a-8123-3d487c93ac6b/barbican-keystone-listener-log/0.log" Oct 10 14:21:50 crc kubenswrapper[4745]: I1010 14:21:50.006735 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-57477cc59-r674m_82ee4ff2-4fa9-42fc-91dd-ee171c443ffa/barbican-worker-log/0.log" Oct 10 14:21:50 crc kubenswrapper[4745]: I1010 14:21:50.018829 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-57477cc59-r674m_82ee4ff2-4fa9-42fc-91dd-ee171c443ffa/barbican-worker/0.log" Oct 10 14:21:50 crc kubenswrapper[4745]: I1010 14:21:50.247814 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-zp2sh_7f174876-f891-4508-97e6-e31624990bc8/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:21:50 crc kubenswrapper[4745]: I1010 14:21:50.272853 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0d791e6b-724b-4631-8a50-98b6dfd108f2/ceilometer-central-agent/0.log" Oct 10 14:21:50 crc kubenswrapper[4745]: I1010 14:21:50.323881 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0d791e6b-724b-4631-8a50-98b6dfd108f2/ceilometer-notification-agent/0.log" Oct 10 14:21:50 crc kubenswrapper[4745]: I1010 14:21:50.423918 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0d791e6b-724b-4631-8a50-98b6dfd108f2/proxy-httpd/0.log" Oct 10 14:21:50 crc kubenswrapper[4745]: I1010 14:21:50.463600 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0d791e6b-724b-4631-8a50-98b6dfd108f2/sg-core/0.log" Oct 10 14:21:50 crc kubenswrapper[4745]: I1010 14:21:50.566961 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_90b61af2-8fcd-40b6-8dc5-92207d219f45/cinder-api/0.log" Oct 10 14:21:50 crc kubenswrapper[4745]: I1010 14:21:50.606854 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_90b61af2-8fcd-40b6-8dc5-92207d219f45/cinder-api-log/0.log" Oct 10 14:21:50 crc kubenswrapper[4745]: I1010 14:21:50.721761 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_513c43ae-c96a-4ef0-8162-708490db52d6/cinder-scheduler/0.log" Oct 10 14:21:50 crc kubenswrapper[4745]: I1010 14:21:50.805770 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_513c43ae-c96a-4ef0-8162-708490db52d6/probe/0.log" Oct 10 14:21:50 crc kubenswrapper[4745]: I1010 14:21:50.922100 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-xprnx_986809df-787f-46ec-b4ac-dc28eaf82cc7/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:21:51 crc kubenswrapper[4745]: I1010 14:21:51.023223 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-2w7mx_0a993d65-ad98-4529-95ad-1663fb206c5a/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:21:51 crc kubenswrapper[4745]: I1010 14:21:51.111237 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-cvz8m_4329eb3b-cd56-417e-9da2-361ab8817091/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:21:51 crc kubenswrapper[4745]: I1010 14:21:51.264901 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-4hmj7_b5d53599-f2c8-4120-b767-72793898c000/init/0.log" Oct 10 14:21:51 crc kubenswrapper[4745]: I1010 14:21:51.517834 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-4hmj7_b5d53599-f2c8-4120-b767-72793898c000/init/0.log" Oct 10 14:21:51 crc kubenswrapper[4745]: I1010 14:21:51.533101 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-cwcr8_e81c2fb9-d588-4ca2-b089-d1a4d1cf9f18/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:21:51 crc kubenswrapper[4745]: I1010 14:21:51.533858 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-4hmj7_b5d53599-f2c8-4120-b767-72793898c000/dnsmasq-dns/0.log" Oct 10 14:21:51 crc kubenswrapper[4745]: I1010 14:21:51.696091 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d2ba148a-0fdd-45d3-91e7-106c98cc7d2e/glance-log/0.log" Oct 10 14:21:51 crc kubenswrapper[4745]: I1010 14:21:51.750200 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d2ba148a-0fdd-45d3-91e7-106c98cc7d2e/glance-httpd/0.log" Oct 10 14:21:51 crc kubenswrapper[4745]: I1010 14:21:51.942530 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0a7607af-7365-45dc-9a8e-91719c20cdcf/glance-httpd/0.log" Oct 10 14:21:51 crc kubenswrapper[4745]: I1010 14:21:51.980064 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0a7607af-7365-45dc-9a8e-91719c20cdcf/glance-log/0.log" Oct 10 14:21:52 crc kubenswrapper[4745]: I1010 14:21:52.066511 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-8574589b48-pc9wr_fb40990e-1db6-4eab-9365-06678441a9ba/horizon/0.log" Oct 10 14:21:52 crc kubenswrapper[4745]: I1010 14:21:52.323707 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-rngcx_8360e97b-9df5-4790-8796-06c6280f323d/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:21:52 crc kubenswrapper[4745]: I1010 14:21:52.444795 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-8574589b48-pc9wr_fb40990e-1db6-4eab-9365-06678441a9ba/horizon-log/0.log" Oct 10 14:21:52 crc kubenswrapper[4745]: I1010 14:21:52.463377 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-hflsg_292290b6-c1e2-4b2b-955b-699aa826627d/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:21:52 crc kubenswrapper[4745]: I1010 14:21:52.676854 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29335081-sq6xb_a5f9c2a5-82c5-456e-b023-667c13b4d328/keystone-cron/0.log" Oct 10 14:21:52 crc kubenswrapper[4745]: I1010 14:21:52.781688 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-745bfbd6-ms6d2_a870bece-0e73-47f8-b0d2-79ee3e3cf637/keystone-api/0.log" Oct 10 14:21:52 crc kubenswrapper[4745]: I1010 14:21:52.920445 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_2c1adab6-47ba-4e16-b607-e61f92db1b53/kube-state-metrics/0.log" Oct 10 14:21:53 crc kubenswrapper[4745]: I1010 14:21:53.004381 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-bb8zg_4709df73-fa0c-4b05-aa37-8d32a4aa52f1/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:21:53 crc kubenswrapper[4745]: I1010 14:21:53.321522 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-86b9f849df-zktvk_668e04d2-c914-4325-acab-c8638172e4cc/neutron-httpd/0.log" Oct 10 14:21:53 crc kubenswrapper[4745]: I1010 14:21:53.426396 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-86b9f849df-zktvk_668e04d2-c914-4325-acab-c8638172e4cc/neutron-api/0.log" Oct 10 14:21:53 crc kubenswrapper[4745]: I1010 14:21:53.433754 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-m8dh2_4674795d-7ada-40eb-93d6-524c7af53ee3/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:21:54 crc kubenswrapper[4745]: I1010 14:21:54.102420 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_b1801b5b-c008-463f-952b-92ddef21faae/nova-api-log/0.log" Oct 10 14:21:54 crc kubenswrapper[4745]: I1010 14:21:54.112581 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_ac183214-b5c6-4e33-8045-ddef84c43323/nova-cell0-conductor-conductor/0.log" Oct 10 14:21:54 crc kubenswrapper[4745]: I1010 14:21:54.349435 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_87ac2f43-7e80-4310-8d3f-5750ae54cc06/nova-cell1-conductor-conductor/0.log" Oct 10 14:21:54 crc kubenswrapper[4745]: I1010 14:21:54.514309 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_6ff9dc11-4059-440b-a21f-8dd6ef566d36/nova-cell1-novncproxy-novncproxy/0.log" Oct 10 14:21:54 crc kubenswrapper[4745]: I1010 14:21:54.523339 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_b1801b5b-c008-463f-952b-92ddef21faae/nova-api-api/0.log" Oct 10 14:21:54 crc kubenswrapper[4745]: I1010 14:21:54.649974 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-nww5s_274ab6cc-4796-493d-ad35-5cf55195151c/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:21:55 crc kubenswrapper[4745]: I1010 14:21:55.031808 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_254ea8d2-b80e-40c0-8e33-3459e73878de/nova-metadata-log/0.log" Oct 10 14:21:55 crc kubenswrapper[4745]: I1010 14:21:55.330816 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c22070ab-5d89-42fa-b04a-640ec3045982/mysql-bootstrap/0.log" Oct 10 14:21:55 crc kubenswrapper[4745]: I1010 14:21:55.355999 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_8d13d787-633b-4ae3-8fbd-013c3de88709/nova-scheduler-scheduler/0.log" Oct 10 14:21:55 crc kubenswrapper[4745]: I1010 14:21:55.527838 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c22070ab-5d89-42fa-b04a-640ec3045982/galera/0.log" Oct 10 14:21:55 crc kubenswrapper[4745]: I1010 14:21:55.586801 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c22070ab-5d89-42fa-b04a-640ec3045982/mysql-bootstrap/0.log" Oct 10 14:21:55 crc kubenswrapper[4745]: I1010 14:21:55.737402 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_35e75435-406e-4053-a1c2-2a7fc992b31e/mysql-bootstrap/0.log" Oct 10 14:21:55 crc kubenswrapper[4745]: I1010 14:21:55.922118 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_35e75435-406e-4053-a1c2-2a7fc992b31e/mysql-bootstrap/0.log" Oct 10 14:21:55 crc kubenswrapper[4745]: I1010 14:21:55.955868 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_35e75435-406e-4053-a1c2-2a7fc992b31e/galera/0.log" Oct 10 14:21:56 crc kubenswrapper[4745]: I1010 14:21:56.114872 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_5cc91ec2-6a94-4923-82cf-d488c077a97b/openstackclient/0.log" Oct 10 14:21:56 crc kubenswrapper[4745]: I1010 14:21:56.212233 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-92pgn_513dcce5-f68e-42a4-ba94-44013577e96e/ovn-controller/0.log" Oct 10 14:21:56 crc kubenswrapper[4745]: I1010 14:21:56.283854 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_254ea8d2-b80e-40c0-8e33-3459e73878de/nova-metadata-metadata/0.log" Oct 10 14:21:56 crc kubenswrapper[4745]: I1010 14:21:56.390260 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-xhz46_3f7ecb0e-ebe5-48f5-a89e-adf1f9fa01ce/openstack-network-exporter/0.log" Oct 10 14:21:56 crc kubenswrapper[4745]: I1010 14:21:56.538713 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s2qdr_e9b37d55-6160-43d6-8957-4cc31c95251d/ovsdb-server-init/0.log" Oct 10 14:21:56 crc kubenswrapper[4745]: I1010 14:21:56.675420 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s2qdr_e9b37d55-6160-43d6-8957-4cc31c95251d/ovs-vswitchd/0.log" Oct 10 14:21:56 crc kubenswrapper[4745]: I1010 14:21:56.684668 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s2qdr_e9b37d55-6160-43d6-8957-4cc31c95251d/ovsdb-server-init/0.log" Oct 10 14:21:56 crc kubenswrapper[4745]: I1010 14:21:56.765005 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s2qdr_e9b37d55-6160-43d6-8957-4cc31c95251d/ovsdb-server/0.log" Oct 10 14:21:56 crc kubenswrapper[4745]: I1010 14:21:56.892952 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-5fz6h_ca7c907d-5166-4e04-b485-04d0b903d6cd/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:21:56 crc kubenswrapper[4745]: I1010 14:21:56.951870 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_6f12c742-f26e-473c-ab59-02217e1b24c6/openstack-network-exporter/0.log" Oct 10 14:21:56 crc kubenswrapper[4745]: I1010 14:21:56.997126 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_6f12c742-f26e-473c-ab59-02217e1b24c6/ovn-northd/0.log" Oct 10 14:21:57 crc kubenswrapper[4745]: I1010 14:21:57.170625 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_dd428c33-2518-41ce-8dd4-b62fe3059525/openstack-network-exporter/0.log" Oct 10 14:21:57 crc kubenswrapper[4745]: I1010 14:21:57.181054 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_dd428c33-2518-41ce-8dd4-b62fe3059525/ovsdbserver-nb/0.log" Oct 10 14:21:57 crc kubenswrapper[4745]: I1010 14:21:57.347115 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_3894eba1-5402-4b7f-9859-a57827e29f2b/openstack-network-exporter/0.log" Oct 10 14:21:57 crc kubenswrapper[4745]: I1010 14:21:57.431153 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_3894eba1-5402-4b7f-9859-a57827e29f2b/ovsdbserver-sb/0.log" Oct 10 14:21:57 crc kubenswrapper[4745]: I1010 14:21:57.513201 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7b8bc58bc6-p44zs_5cf84964-c89c-4990-8320-823b10461186/placement-api/0.log" Oct 10 14:21:57 crc kubenswrapper[4745]: I1010 14:21:57.623454 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7b8bc58bc6-p44zs_5cf84964-c89c-4990-8320-823b10461186/placement-log/0.log" Oct 10 14:21:57 crc kubenswrapper[4745]: I1010 14:21:57.697670 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_972f7a23-bb59-4b6c-8ee7-a51647bbc261/setup-container/0.log" Oct 10 14:21:57 crc kubenswrapper[4745]: I1010 14:21:57.882329 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_972f7a23-bb59-4b6c-8ee7-a51647bbc261/rabbitmq/0.log" Oct 10 14:21:57 crc kubenswrapper[4745]: I1010 14:21:57.948512 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3f7d0aba-ccad-4d20-9ab1-cf7d080571c1/setup-container/0.log" Oct 10 14:21:57 crc kubenswrapper[4745]: I1010 14:21:57.962851 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_972f7a23-bb59-4b6c-8ee7-a51647bbc261/setup-container/0.log" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.159457 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3f7d0aba-ccad-4d20-9ab1-cf7d080571c1/rabbitmq/0.log" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.162331 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-crm9n_b812bcc0-a0b6-4ef6-afd2-d0b8234f1ab4/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.175523 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3f7d0aba-ccad-4d20-9ab1-cf7d080571c1/setup-container/0.log" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.223976 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zjf6z"] Oct 10 14:21:58 crc kubenswrapper[4745]: E1010 14:21:58.224379 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b69b971f-6f8a-4486-abe7-9052b04587fd" containerName="container-00" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.224396 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="b69b971f-6f8a-4486-abe7-9052b04587fd" containerName="container-00" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.224610 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="b69b971f-6f8a-4486-abe7-9052b04587fd" containerName="container-00" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.225918 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zjf6z" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.239148 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zjf6z"] Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.280827 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3a02448-64e8-4add-8561-305ab7ad2e19-utilities\") pod \"redhat-marketplace-zjf6z\" (UID: \"f3a02448-64e8-4add-8561-305ab7ad2e19\") " pod="openshift-marketplace/redhat-marketplace-zjf6z" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.280877 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c9hr\" (UniqueName: \"kubernetes.io/projected/f3a02448-64e8-4add-8561-305ab7ad2e19-kube-api-access-4c9hr\") pod \"redhat-marketplace-zjf6z\" (UID: \"f3a02448-64e8-4add-8561-305ab7ad2e19\") " pod="openshift-marketplace/redhat-marketplace-zjf6z" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.280905 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3a02448-64e8-4add-8561-305ab7ad2e19-catalog-content\") pod \"redhat-marketplace-zjf6z\" (UID: \"f3a02448-64e8-4add-8561-305ab7ad2e19\") " pod="openshift-marketplace/redhat-marketplace-zjf6z" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.382490 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3a02448-64e8-4add-8561-305ab7ad2e19-utilities\") pod \"redhat-marketplace-zjf6z\" (UID: \"f3a02448-64e8-4add-8561-305ab7ad2e19\") " pod="openshift-marketplace/redhat-marketplace-zjf6z" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.382543 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c9hr\" (UniqueName: \"kubernetes.io/projected/f3a02448-64e8-4add-8561-305ab7ad2e19-kube-api-access-4c9hr\") pod \"redhat-marketplace-zjf6z\" (UID: \"f3a02448-64e8-4add-8561-305ab7ad2e19\") " pod="openshift-marketplace/redhat-marketplace-zjf6z" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.382568 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3a02448-64e8-4add-8561-305ab7ad2e19-catalog-content\") pod \"redhat-marketplace-zjf6z\" (UID: \"f3a02448-64e8-4add-8561-305ab7ad2e19\") " pod="openshift-marketplace/redhat-marketplace-zjf6z" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.383077 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3a02448-64e8-4add-8561-305ab7ad2e19-catalog-content\") pod \"redhat-marketplace-zjf6z\" (UID: \"f3a02448-64e8-4add-8561-305ab7ad2e19\") " pod="openshift-marketplace/redhat-marketplace-zjf6z" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.384379 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3a02448-64e8-4add-8561-305ab7ad2e19-utilities\") pod \"redhat-marketplace-zjf6z\" (UID: \"f3a02448-64e8-4add-8561-305ab7ad2e19\") " pod="openshift-marketplace/redhat-marketplace-zjf6z" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.420555 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c9hr\" (UniqueName: \"kubernetes.io/projected/f3a02448-64e8-4add-8561-305ab7ad2e19-kube-api-access-4c9hr\") pod \"redhat-marketplace-zjf6z\" (UID: \"f3a02448-64e8-4add-8561-305ab7ad2e19\") " pod="openshift-marketplace/redhat-marketplace-zjf6z" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.441250 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-fvppw_4fda6419-bf48-4f0e-8ff8-da6b9c2da25f/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.603317 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zjf6z" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.644895 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-b5fc9_167c6043-6db1-4ead-8cf4-4040842b8d88/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.889432 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-vr2vp_29cef4fe-f039-411b-a979-91362f56cfd7/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:21:58 crc kubenswrapper[4745]: I1010 14:21:58.891670 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-2mslp_0d120415-3123-4b2b-a29d-b1b9c822ed7c/ssh-known-hosts-edpm-deployment/0.log" Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.095283 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zjf6z"] Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.130432 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5d77644dd7-jv6wz_6aa7ade8-3d4b-4168-80a9-673c73b919a3/proxy-server/0.log" Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.227341 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5d77644dd7-jv6wz_6aa7ade8-3d4b-4168-80a9-673c73b919a3/proxy-httpd/0.log" Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.281507 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-8kzzn_749bd815-8704-47d6-b5a0-4e129ab976c1/swift-ring-rebalance/0.log" Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.339149 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/account-auditor/0.log" Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.454344 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/account-reaper/0.log" Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.571066 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/account-server/0.log" Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.598026 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/account-replicator/0.log" Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.700531 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/container-auditor/0.log" Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.720190 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/container-replicator/0.log" Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.746362 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/container-server/0.log" Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.823169 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/container-updater/0.log" Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.845682 4745 generic.go:334] "Generic (PLEG): container finished" podID="f3a02448-64e8-4add-8561-305ab7ad2e19" containerID="422b3041f41e5498677a9972cd92ccfdf07c5425902a539dbc7ae88f7ca32ccd" exitCode=0 Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.845746 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjf6z" event={"ID":"f3a02448-64e8-4add-8561-305ab7ad2e19","Type":"ContainerDied","Data":"422b3041f41e5498677a9972cd92ccfdf07c5425902a539dbc7ae88f7ca32ccd"} Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.845808 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjf6z" event={"ID":"f3a02448-64e8-4add-8561-305ab7ad2e19","Type":"ContainerStarted","Data":"da9e150a5034a255383a8ce5c97463855a5c91f1b146b80bf852aeabec9767aa"} Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.848276 4745 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.955260 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/object-auditor/0.log" Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.959463 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/object-replicator/0.log" Oct 10 14:21:59 crc kubenswrapper[4745]: I1010 14:21:59.973238 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/object-expirer/0.log" Oct 10 14:22:00 crc kubenswrapper[4745]: I1010 14:22:00.020987 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/object-server/0.log" Oct 10 14:22:00 crc kubenswrapper[4745]: I1010 14:22:00.149507 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/swift-recon-cron/0.log" Oct 10 14:22:00 crc kubenswrapper[4745]: I1010 14:22:00.158692 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/rsync/0.log" Oct 10 14:22:00 crc kubenswrapper[4745]: I1010 14:22:00.183036 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_00893894-3690-4a11-a2f7-0d1dc6c14e86/object-updater/0.log" Oct 10 14:22:00 crc kubenswrapper[4745]: I1010 14:22:00.447335 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_55dab8c7-31e2-4694-9f9e-a3272376cddd/tempest-tests-tempest-tests-runner/0.log" Oct 10 14:22:00 crc kubenswrapper[4745]: I1010 14:22:00.465763 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-s7pz7_9c27594e-14e8-4b37-860c-19d4c47605cc/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:22:00 crc kubenswrapper[4745]: I1010 14:22:00.633397 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_cb7b00b8-301c-4b88-a189-2019e53366ec/test-operator-logs-container/0.log" Oct 10 14:22:00 crc kubenswrapper[4745]: I1010 14:22:00.713188 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-b9582_e4f15e34-6886-448d-9ed1-7a5cbaf0fda2/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 14:22:00 crc kubenswrapper[4745]: I1010 14:22:00.855030 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjf6z" event={"ID":"f3a02448-64e8-4add-8561-305ab7ad2e19","Type":"ContainerStarted","Data":"0023e10ebad95847d0e565b585fc9c196cc0877616deed52952ec692c54ead4c"} Oct 10 14:22:01 crc kubenswrapper[4745]: I1010 14:22:01.863213 4745 generic.go:334] "Generic (PLEG): container finished" podID="f3a02448-64e8-4add-8561-305ab7ad2e19" containerID="0023e10ebad95847d0e565b585fc9c196cc0877616deed52952ec692c54ead4c" exitCode=0 Oct 10 14:22:01 crc kubenswrapper[4745]: I1010 14:22:01.864232 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjf6z" event={"ID":"f3a02448-64e8-4add-8561-305ab7ad2e19","Type":"ContainerDied","Data":"0023e10ebad95847d0e565b585fc9c196cc0877616deed52952ec692c54ead4c"} Oct 10 14:22:02 crc kubenswrapper[4745]: I1010 14:22:02.873600 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjf6z" event={"ID":"f3a02448-64e8-4add-8561-305ab7ad2e19","Type":"ContainerStarted","Data":"93d0ee2fc5a19356666c4ef9acc527fef69bc74db3b774ac70ac08452b824c0c"} Oct 10 14:22:02 crc kubenswrapper[4745]: I1010 14:22:02.918986 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zjf6z" podStartSLOduration=2.482601454 podStartE2EDuration="4.918961593s" podCreationTimestamp="2025-10-10 14:21:58 +0000 UTC" firstStartedPulling="2025-10-10 14:21:59.848008015 +0000 UTC m=+3833.745664778" lastFinishedPulling="2025-10-10 14:22:02.284368154 +0000 UTC m=+3836.182024917" observedRunningTime="2025-10-10 14:22:02.9070062 +0000 UTC m=+3836.804662953" watchObservedRunningTime="2025-10-10 14:22:02.918961593 +0000 UTC m=+3836.816618356" Oct 10 14:22:08 crc kubenswrapper[4745]: I1010 14:22:08.603430 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zjf6z" Oct 10 14:22:08 crc kubenswrapper[4745]: I1010 14:22:08.603985 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zjf6z" Oct 10 14:22:08 crc kubenswrapper[4745]: I1010 14:22:08.654072 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zjf6z" Oct 10 14:22:09 crc kubenswrapper[4745]: I1010 14:22:09.001290 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zjf6z" Oct 10 14:22:09 crc kubenswrapper[4745]: I1010 14:22:09.045391 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zjf6z"] Oct 10 14:22:09 crc kubenswrapper[4745]: I1010 14:22:09.320685 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_9e79afba-7f24-4e67-9c37-973256aac339/memcached/0.log" Oct 10 14:22:10 crc kubenswrapper[4745]: I1010 14:22:10.959779 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zjf6z" podUID="f3a02448-64e8-4add-8561-305ab7ad2e19" containerName="registry-server" containerID="cri-o://93d0ee2fc5a19356666c4ef9acc527fef69bc74db3b774ac70ac08452b824c0c" gracePeriod=2 Oct 10 14:22:11 crc kubenswrapper[4745]: I1010 14:22:11.411844 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zjf6z" Oct 10 14:22:11 crc kubenswrapper[4745]: I1010 14:22:11.572801 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3a02448-64e8-4add-8561-305ab7ad2e19-catalog-content\") pod \"f3a02448-64e8-4add-8561-305ab7ad2e19\" (UID: \"f3a02448-64e8-4add-8561-305ab7ad2e19\") " Oct 10 14:22:11 crc kubenswrapper[4745]: I1010 14:22:11.572951 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3a02448-64e8-4add-8561-305ab7ad2e19-utilities\") pod \"f3a02448-64e8-4add-8561-305ab7ad2e19\" (UID: \"f3a02448-64e8-4add-8561-305ab7ad2e19\") " Oct 10 14:22:11 crc kubenswrapper[4745]: I1010 14:22:11.573273 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4c9hr\" (UniqueName: \"kubernetes.io/projected/f3a02448-64e8-4add-8561-305ab7ad2e19-kube-api-access-4c9hr\") pod \"f3a02448-64e8-4add-8561-305ab7ad2e19\" (UID: \"f3a02448-64e8-4add-8561-305ab7ad2e19\") " Oct 10 14:22:11 crc kubenswrapper[4745]: I1010 14:22:11.573600 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3a02448-64e8-4add-8561-305ab7ad2e19-utilities" (OuterVolumeSpecName: "utilities") pod "f3a02448-64e8-4add-8561-305ab7ad2e19" (UID: "f3a02448-64e8-4add-8561-305ab7ad2e19"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:22:11 crc kubenswrapper[4745]: I1010 14:22:11.573903 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3a02448-64e8-4add-8561-305ab7ad2e19-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:22:11 crc kubenswrapper[4745]: I1010 14:22:11.578818 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3a02448-64e8-4add-8561-305ab7ad2e19-kube-api-access-4c9hr" (OuterVolumeSpecName: "kube-api-access-4c9hr") pod "f3a02448-64e8-4add-8561-305ab7ad2e19" (UID: "f3a02448-64e8-4add-8561-305ab7ad2e19"). InnerVolumeSpecName "kube-api-access-4c9hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:22:11 crc kubenswrapper[4745]: I1010 14:22:11.589462 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3a02448-64e8-4add-8561-305ab7ad2e19-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3a02448-64e8-4add-8561-305ab7ad2e19" (UID: "f3a02448-64e8-4add-8561-305ab7ad2e19"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:22:11 crc kubenswrapper[4745]: I1010 14:22:11.675635 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3a02448-64e8-4add-8561-305ab7ad2e19-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:22:11 crc kubenswrapper[4745]: I1010 14:22:11.675669 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4c9hr\" (UniqueName: \"kubernetes.io/projected/f3a02448-64e8-4add-8561-305ab7ad2e19-kube-api-access-4c9hr\") on node \"crc\" DevicePath \"\"" Oct 10 14:22:11 crc kubenswrapper[4745]: I1010 14:22:11.970086 4745 generic.go:334] "Generic (PLEG): container finished" podID="f3a02448-64e8-4add-8561-305ab7ad2e19" containerID="93d0ee2fc5a19356666c4ef9acc527fef69bc74db3b774ac70ac08452b824c0c" exitCode=0 Oct 10 14:22:11 crc kubenswrapper[4745]: I1010 14:22:11.970156 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zjf6z" Oct 10 14:22:11 crc kubenswrapper[4745]: I1010 14:22:11.970187 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjf6z" event={"ID":"f3a02448-64e8-4add-8561-305ab7ad2e19","Type":"ContainerDied","Data":"93d0ee2fc5a19356666c4ef9acc527fef69bc74db3b774ac70ac08452b824c0c"} Oct 10 14:22:11 crc kubenswrapper[4745]: I1010 14:22:11.970475 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjf6z" event={"ID":"f3a02448-64e8-4add-8561-305ab7ad2e19","Type":"ContainerDied","Data":"da9e150a5034a255383a8ce5c97463855a5c91f1b146b80bf852aeabec9767aa"} Oct 10 14:22:11 crc kubenswrapper[4745]: I1010 14:22:11.970502 4745 scope.go:117] "RemoveContainer" containerID="93d0ee2fc5a19356666c4ef9acc527fef69bc74db3b774ac70ac08452b824c0c" Oct 10 14:22:11 crc kubenswrapper[4745]: I1010 14:22:11.991888 4745 scope.go:117] "RemoveContainer" containerID="0023e10ebad95847d0e565b585fc9c196cc0877616deed52952ec692c54ead4c" Oct 10 14:22:12 crc kubenswrapper[4745]: I1010 14:22:12.008773 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zjf6z"] Oct 10 14:22:12 crc kubenswrapper[4745]: I1010 14:22:12.022292 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zjf6z"] Oct 10 14:22:12 crc kubenswrapper[4745]: I1010 14:22:12.022371 4745 scope.go:117] "RemoveContainer" containerID="422b3041f41e5498677a9972cd92ccfdf07c5425902a539dbc7ae88f7ca32ccd" Oct 10 14:22:12 crc kubenswrapper[4745]: I1010 14:22:12.070447 4745 scope.go:117] "RemoveContainer" containerID="93d0ee2fc5a19356666c4ef9acc527fef69bc74db3b774ac70ac08452b824c0c" Oct 10 14:22:12 crc kubenswrapper[4745]: E1010 14:22:12.070912 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93d0ee2fc5a19356666c4ef9acc527fef69bc74db3b774ac70ac08452b824c0c\": container with ID starting with 93d0ee2fc5a19356666c4ef9acc527fef69bc74db3b774ac70ac08452b824c0c not found: ID does not exist" containerID="93d0ee2fc5a19356666c4ef9acc527fef69bc74db3b774ac70ac08452b824c0c" Oct 10 14:22:12 crc kubenswrapper[4745]: I1010 14:22:12.070960 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93d0ee2fc5a19356666c4ef9acc527fef69bc74db3b774ac70ac08452b824c0c"} err="failed to get container status \"93d0ee2fc5a19356666c4ef9acc527fef69bc74db3b774ac70ac08452b824c0c\": rpc error: code = NotFound desc = could not find container \"93d0ee2fc5a19356666c4ef9acc527fef69bc74db3b774ac70ac08452b824c0c\": container with ID starting with 93d0ee2fc5a19356666c4ef9acc527fef69bc74db3b774ac70ac08452b824c0c not found: ID does not exist" Oct 10 14:22:12 crc kubenswrapper[4745]: I1010 14:22:12.070988 4745 scope.go:117] "RemoveContainer" containerID="0023e10ebad95847d0e565b585fc9c196cc0877616deed52952ec692c54ead4c" Oct 10 14:22:12 crc kubenswrapper[4745]: E1010 14:22:12.071464 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0023e10ebad95847d0e565b585fc9c196cc0877616deed52952ec692c54ead4c\": container with ID starting with 0023e10ebad95847d0e565b585fc9c196cc0877616deed52952ec692c54ead4c not found: ID does not exist" containerID="0023e10ebad95847d0e565b585fc9c196cc0877616deed52952ec692c54ead4c" Oct 10 14:22:12 crc kubenswrapper[4745]: I1010 14:22:12.071499 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0023e10ebad95847d0e565b585fc9c196cc0877616deed52952ec692c54ead4c"} err="failed to get container status \"0023e10ebad95847d0e565b585fc9c196cc0877616deed52952ec692c54ead4c\": rpc error: code = NotFound desc = could not find container \"0023e10ebad95847d0e565b585fc9c196cc0877616deed52952ec692c54ead4c\": container with ID starting with 0023e10ebad95847d0e565b585fc9c196cc0877616deed52952ec692c54ead4c not found: ID does not exist" Oct 10 14:22:12 crc kubenswrapper[4745]: I1010 14:22:12.071523 4745 scope.go:117] "RemoveContainer" containerID="422b3041f41e5498677a9972cd92ccfdf07c5425902a539dbc7ae88f7ca32ccd" Oct 10 14:22:12 crc kubenswrapper[4745]: E1010 14:22:12.072462 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"422b3041f41e5498677a9972cd92ccfdf07c5425902a539dbc7ae88f7ca32ccd\": container with ID starting with 422b3041f41e5498677a9972cd92ccfdf07c5425902a539dbc7ae88f7ca32ccd not found: ID does not exist" containerID="422b3041f41e5498677a9972cd92ccfdf07c5425902a539dbc7ae88f7ca32ccd" Oct 10 14:22:12 crc kubenswrapper[4745]: I1010 14:22:12.072506 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"422b3041f41e5498677a9972cd92ccfdf07c5425902a539dbc7ae88f7ca32ccd"} err="failed to get container status \"422b3041f41e5498677a9972cd92ccfdf07c5425902a539dbc7ae88f7ca32ccd\": rpc error: code = NotFound desc = could not find container \"422b3041f41e5498677a9972cd92ccfdf07c5425902a539dbc7ae88f7ca32ccd\": container with ID starting with 422b3041f41e5498677a9972cd92ccfdf07c5425902a539dbc7ae88f7ca32ccd not found: ID does not exist" Oct 10 14:22:12 crc kubenswrapper[4745]: I1010 14:22:12.755255 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3a02448-64e8-4add-8561-305ab7ad2e19" path="/var/lib/kubelet/pods/f3a02448-64e8-4add-8561-305ab7ad2e19/volumes" Oct 10 14:22:22 crc kubenswrapper[4745]: I1010 14:22:22.707706 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc_128069b5-6f89-4da5-9e35-bfe14f1ebd50/util/0.log" Oct 10 14:22:22 crc kubenswrapper[4745]: I1010 14:22:22.843188 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc_128069b5-6f89-4da5-9e35-bfe14f1ebd50/util/0.log" Oct 10 14:22:22 crc kubenswrapper[4745]: I1010 14:22:22.870821 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc_128069b5-6f89-4da5-9e35-bfe14f1ebd50/pull/0.log" Oct 10 14:22:22 crc kubenswrapper[4745]: I1010 14:22:22.882050 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc_128069b5-6f89-4da5-9e35-bfe14f1ebd50/pull/0.log" Oct 10 14:22:23 crc kubenswrapper[4745]: I1010 14:22:23.066492 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc_128069b5-6f89-4da5-9e35-bfe14f1ebd50/pull/0.log" Oct 10 14:22:23 crc kubenswrapper[4745]: I1010 14:22:23.072252 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc_128069b5-6f89-4da5-9e35-bfe14f1ebd50/util/0.log" Oct 10 14:22:23 crc kubenswrapper[4745]: I1010 14:22:23.100462 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_73cb35fc92d4cf7f2c3e926149d6071361115ad32d230d693a9faf4168lghcc_128069b5-6f89-4da5-9e35-bfe14f1ebd50/extract/0.log" Oct 10 14:22:23 crc kubenswrapper[4745]: I1010 14:22:23.222209 4745 scope.go:117] "RemoveContainer" containerID="52dfac7e7ddc8facd62fb7de45ce63d37600f743df740eb3a4d7505170bdac35" Oct 10 14:22:23 crc kubenswrapper[4745]: I1010 14:22:23.248597 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-htmcv_730b5555-5f6d-4e16-91e0-faacad56d777/kube-rbac-proxy/0.log" Oct 10 14:22:23 crc kubenswrapper[4745]: I1010 14:22:23.310656 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-htmcv_730b5555-5f6d-4e16-91e0-faacad56d777/manager/0.log" Oct 10 14:22:23 crc kubenswrapper[4745]: I1010 14:22:23.329770 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-c7986_5ca519af-249e-48f0-8669-77bd548be8f3/kube-rbac-proxy/0.log" Oct 10 14:22:23 crc kubenswrapper[4745]: I1010 14:22:23.487810 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-c7986_5ca519af-249e-48f0-8669-77bd548be8f3/manager/0.log" Oct 10 14:22:23 crc kubenswrapper[4745]: I1010 14:22:23.514201 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-fpcxk_dff847b8-915b-4b64-85c3-d7c20d1282a0/kube-rbac-proxy/0.log" Oct 10 14:22:23 crc kubenswrapper[4745]: I1010 14:22:23.516137 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-fpcxk_dff847b8-915b-4b64-85c3-d7c20d1282a0/manager/0.log" Oct 10 14:22:23 crc kubenswrapper[4745]: I1010 14:22:23.712611 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-tz665_446e0b63-81e5-47fb-b110-cfa5393e5311/kube-rbac-proxy/0.log" Oct 10 14:22:23 crc kubenswrapper[4745]: I1010 14:22:23.760747 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-tz665_446e0b63-81e5-47fb-b110-cfa5393e5311/manager/0.log" Oct 10 14:22:23 crc kubenswrapper[4745]: I1010 14:22:23.891186 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-2w6nw_7f07678c-0cc2-47a5-b244-72ae3aa55baa/manager/0.log" Oct 10 14:22:23 crc kubenswrapper[4745]: I1010 14:22:23.900474 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-2w6nw_7f07678c-0cc2-47a5-b244-72ae3aa55baa/kube-rbac-proxy/0.log" Oct 10 14:22:23 crc kubenswrapper[4745]: I1010 14:22:23.918421 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-qmbvh_1d2b124e-5495-42bd-b675-4eed70ccc446/kube-rbac-proxy/0.log" Oct 10 14:22:24 crc kubenswrapper[4745]: I1010 14:22:24.044815 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-qmbvh_1d2b124e-5495-42bd-b675-4eed70ccc446/manager/0.log" Oct 10 14:22:24 crc kubenswrapper[4745]: I1010 14:22:24.089687 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-dt9kx_443416f7-e4ce-41f2-b546-9344f0cb72a7/kube-rbac-proxy/0.log" Oct 10 14:22:24 crc kubenswrapper[4745]: I1010 14:22:24.274286 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-dt9kx_443416f7-e4ce-41f2-b546-9344f0cb72a7/manager/0.log" Oct 10 14:22:24 crc kubenswrapper[4745]: I1010 14:22:24.344288 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-d5qnn_9fdf7ed4-1e18-43c0-b305-2b9b226f4634/kube-rbac-proxy/0.log" Oct 10 14:22:24 crc kubenswrapper[4745]: I1010 14:22:24.357585 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-d5qnn_9fdf7ed4-1e18-43c0-b305-2b9b226f4634/manager/0.log" Oct 10 14:22:24 crc kubenswrapper[4745]: I1010 14:22:24.474837 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-h2bt5_10978ce4-0034-4ad4-9a47-2f1931c09655/kube-rbac-proxy/0.log" Oct 10 14:22:24 crc kubenswrapper[4745]: I1010 14:22:24.631940 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-t9wqw_fc401fc1-e03b-4b32-8782-15e9c2a70e8a/kube-rbac-proxy/0.log" Oct 10 14:22:24 crc kubenswrapper[4745]: I1010 14:22:24.641246 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-h2bt5_10978ce4-0034-4ad4-9a47-2f1931c09655/manager/0.log" Oct 10 14:22:24 crc kubenswrapper[4745]: I1010 14:22:24.660022 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-t9wqw_fc401fc1-e03b-4b32-8782-15e9c2a70e8a/manager/0.log" Oct 10 14:22:24 crc kubenswrapper[4745]: I1010 14:22:24.807878 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-vc9rk_420fdd11-0131-479d-9f7a-47f091cacd83/kube-rbac-proxy/0.log" Oct 10 14:22:24 crc kubenswrapper[4745]: I1010 14:22:24.858427 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-vc9rk_420fdd11-0131-479d-9f7a-47f091cacd83/manager/0.log" Oct 10 14:22:24 crc kubenswrapper[4745]: I1010 14:22:24.971920 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-2dmpc_83b0fe55-090f-448e-95c2-1d3cbbc1b54c/kube-rbac-proxy/0.log" Oct 10 14:22:25 crc kubenswrapper[4745]: I1010 14:22:25.011312 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-2dmpc_83b0fe55-090f-448e-95c2-1d3cbbc1b54c/manager/0.log" Oct 10 14:22:25 crc kubenswrapper[4745]: I1010 14:22:25.018805 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-89lzv_93defbbc-c732-4743-bcff-a1d7f06aefa2/kube-rbac-proxy/0.log" Oct 10 14:22:25 crc kubenswrapper[4745]: I1010 14:22:25.191937 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-xxvtr_e1c2fd77-699a-4c79-9c54-9251c5ad7d17/kube-rbac-proxy/0.log" Oct 10 14:22:25 crc kubenswrapper[4745]: I1010 14:22:25.212109 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-89lzv_93defbbc-c732-4743-bcff-a1d7f06aefa2/manager/0.log" Oct 10 14:22:25 crc kubenswrapper[4745]: I1010 14:22:25.270842 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-xxvtr_e1c2fd77-699a-4c79-9c54-9251c5ad7d17/manager/0.log" Oct 10 14:22:25 crc kubenswrapper[4745]: I1010 14:22:25.406520 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw_0174cdbc-06e0-423f-af63-5467737e6ec5/kube-rbac-proxy/0.log" Oct 10 14:22:25 crc kubenswrapper[4745]: I1010 14:22:25.449342 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757d5jdpw_0174cdbc-06e0-423f-af63-5467737e6ec5/manager/0.log" Oct 10 14:22:25 crc kubenswrapper[4745]: I1010 14:22:25.578253 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6f95c9794b-jtlvc_fd4f5298-2862-4650-950d-abd7f8195942/kube-rbac-proxy/0.log" Oct 10 14:22:25 crc kubenswrapper[4745]: I1010 14:22:25.707816 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7f65679b66-d82sn_b0e08ed6-f210-41b7-907e-d6722025f43e/kube-rbac-proxy/0.log" Oct 10 14:22:25 crc kubenswrapper[4745]: I1010 14:22:25.894425 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-6tszb_40d5edac-be93-424b-8f39-1609f8b7cef4/registry-server/0.log" Oct 10 14:22:25 crc kubenswrapper[4745]: I1010 14:22:25.925542 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7f65679b66-d82sn_b0e08ed6-f210-41b7-907e-d6722025f43e/operator/0.log" Oct 10 14:22:26 crc kubenswrapper[4745]: I1010 14:22:26.081060 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-rz48f_0ec7257b-b487-453c-88e8-d279184acdfc/kube-rbac-proxy/0.log" Oct 10 14:22:26 crc kubenswrapper[4745]: I1010 14:22:26.154303 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-rz48f_0ec7257b-b487-453c-88e8-d279184acdfc/manager/0.log" Oct 10 14:22:26 crc kubenswrapper[4745]: I1010 14:22:26.304444 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-b7w5k_36db8827-b9af-4b9b-87aa-3cfa7d005ee0/manager/0.log" Oct 10 14:22:26 crc kubenswrapper[4745]: I1010 14:22:26.305314 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-b7w5k_36db8827-b9af-4b9b-87aa-3cfa7d005ee0/kube-rbac-proxy/0.log" Oct 10 14:22:26 crc kubenswrapper[4745]: I1010 14:22:26.397087 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-c9gmp_ebc60df2-d5b0-418f-a8cd-e8440706d3be/operator/0.log" Oct 10 14:22:26 crc kubenswrapper[4745]: I1010 14:22:26.536952 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-qzh4v_e6828db3-0a2c-4b98-8cbf-1eba04f13bc5/kube-rbac-proxy/0.log" Oct 10 14:22:26 crc kubenswrapper[4745]: I1010 14:22:26.591624 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-qzh4v_e6828db3-0a2c-4b98-8cbf-1eba04f13bc5/manager/0.log" Oct 10 14:22:26 crc kubenswrapper[4745]: I1010 14:22:26.619110 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6f95c9794b-jtlvc_fd4f5298-2862-4650-950d-abd7f8195942/manager/0.log" Oct 10 14:22:26 crc kubenswrapper[4745]: I1010 14:22:26.754999 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-h4jn2_a5094dca-dde7-4144-bfbc-b4f90c170a21/kube-rbac-proxy/0.log" Oct 10 14:22:26 crc kubenswrapper[4745]: I1010 14:22:26.792885 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-h4jn2_a5094dca-dde7-4144-bfbc-b4f90c170a21/manager/0.log" Oct 10 14:22:26 crc kubenswrapper[4745]: I1010 14:22:26.828346 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-lr987_ba76d6db-10cc-4d7d-b620-e37533730395/manager/0.log" Oct 10 14:22:26 crc kubenswrapper[4745]: I1010 14:22:26.831843 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-lr987_ba76d6db-10cc-4d7d-b620-e37533730395/kube-rbac-proxy/0.log" Oct 10 14:22:26 crc kubenswrapper[4745]: I1010 14:22:26.966405 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-96mn7_6d02fe8c-f0b4-4ef1-8e57-569a836708cd/manager/0.log" Oct 10 14:22:26 crc kubenswrapper[4745]: I1010 14:22:26.993965 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-96mn7_6d02fe8c-f0b4-4ef1-8e57-569a836708cd/kube-rbac-proxy/0.log" Oct 10 14:22:39 crc kubenswrapper[4745]: I1010 14:22:39.479947 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-phgcv"] Oct 10 14:22:39 crc kubenswrapper[4745]: E1010 14:22:39.480878 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3a02448-64e8-4add-8561-305ab7ad2e19" containerName="registry-server" Oct 10 14:22:39 crc kubenswrapper[4745]: I1010 14:22:39.480895 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3a02448-64e8-4add-8561-305ab7ad2e19" containerName="registry-server" Oct 10 14:22:39 crc kubenswrapper[4745]: E1010 14:22:39.480913 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3a02448-64e8-4add-8561-305ab7ad2e19" containerName="extract-utilities" Oct 10 14:22:39 crc kubenswrapper[4745]: I1010 14:22:39.480922 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3a02448-64e8-4add-8561-305ab7ad2e19" containerName="extract-utilities" Oct 10 14:22:39 crc kubenswrapper[4745]: E1010 14:22:39.480952 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3a02448-64e8-4add-8561-305ab7ad2e19" containerName="extract-content" Oct 10 14:22:39 crc kubenswrapper[4745]: I1010 14:22:39.480962 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3a02448-64e8-4add-8561-305ab7ad2e19" containerName="extract-content" Oct 10 14:22:39 crc kubenswrapper[4745]: I1010 14:22:39.481245 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3a02448-64e8-4add-8561-305ab7ad2e19" containerName="registry-server" Oct 10 14:22:39 crc kubenswrapper[4745]: I1010 14:22:39.483016 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-phgcv" Oct 10 14:22:39 crc kubenswrapper[4745]: I1010 14:22:39.491011 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-phgcv"] Oct 10 14:22:39 crc kubenswrapper[4745]: I1010 14:22:39.617434 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91a3691-d1e1-4982-9d4e-b674efda86b1-catalog-content\") pod \"redhat-operators-phgcv\" (UID: \"e91a3691-d1e1-4982-9d4e-b674efda86b1\") " pod="openshift-marketplace/redhat-operators-phgcv" Oct 10 14:22:39 crc kubenswrapper[4745]: I1010 14:22:39.617532 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2ksj\" (UniqueName: \"kubernetes.io/projected/e91a3691-d1e1-4982-9d4e-b674efda86b1-kube-api-access-b2ksj\") pod \"redhat-operators-phgcv\" (UID: \"e91a3691-d1e1-4982-9d4e-b674efda86b1\") " pod="openshift-marketplace/redhat-operators-phgcv" Oct 10 14:22:39 crc kubenswrapper[4745]: I1010 14:22:39.617842 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91a3691-d1e1-4982-9d4e-b674efda86b1-utilities\") pod \"redhat-operators-phgcv\" (UID: \"e91a3691-d1e1-4982-9d4e-b674efda86b1\") " pod="openshift-marketplace/redhat-operators-phgcv" Oct 10 14:22:39 crc kubenswrapper[4745]: I1010 14:22:39.719234 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91a3691-d1e1-4982-9d4e-b674efda86b1-utilities\") pod \"redhat-operators-phgcv\" (UID: \"e91a3691-d1e1-4982-9d4e-b674efda86b1\") " pod="openshift-marketplace/redhat-operators-phgcv" Oct 10 14:22:39 crc kubenswrapper[4745]: I1010 14:22:39.719344 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91a3691-d1e1-4982-9d4e-b674efda86b1-catalog-content\") pod \"redhat-operators-phgcv\" (UID: \"e91a3691-d1e1-4982-9d4e-b674efda86b1\") " pod="openshift-marketplace/redhat-operators-phgcv" Oct 10 14:22:39 crc kubenswrapper[4745]: I1010 14:22:39.719392 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2ksj\" (UniqueName: \"kubernetes.io/projected/e91a3691-d1e1-4982-9d4e-b674efda86b1-kube-api-access-b2ksj\") pod \"redhat-operators-phgcv\" (UID: \"e91a3691-d1e1-4982-9d4e-b674efda86b1\") " pod="openshift-marketplace/redhat-operators-phgcv" Oct 10 14:22:39 crc kubenswrapper[4745]: I1010 14:22:39.719759 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91a3691-d1e1-4982-9d4e-b674efda86b1-utilities\") pod \"redhat-operators-phgcv\" (UID: \"e91a3691-d1e1-4982-9d4e-b674efda86b1\") " pod="openshift-marketplace/redhat-operators-phgcv" Oct 10 14:22:39 crc kubenswrapper[4745]: I1010 14:22:39.719924 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91a3691-d1e1-4982-9d4e-b674efda86b1-catalog-content\") pod \"redhat-operators-phgcv\" (UID: \"e91a3691-d1e1-4982-9d4e-b674efda86b1\") " pod="openshift-marketplace/redhat-operators-phgcv" Oct 10 14:22:39 crc kubenswrapper[4745]: I1010 14:22:39.737979 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2ksj\" (UniqueName: \"kubernetes.io/projected/e91a3691-d1e1-4982-9d4e-b674efda86b1-kube-api-access-b2ksj\") pod \"redhat-operators-phgcv\" (UID: \"e91a3691-d1e1-4982-9d4e-b674efda86b1\") " pod="openshift-marketplace/redhat-operators-phgcv" Oct 10 14:22:39 crc kubenswrapper[4745]: I1010 14:22:39.833408 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-phgcv" Oct 10 14:22:40 crc kubenswrapper[4745]: I1010 14:22:40.292402 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-phgcv"] Oct 10 14:22:41 crc kubenswrapper[4745]: I1010 14:22:41.220512 4745 generic.go:334] "Generic (PLEG): container finished" podID="e91a3691-d1e1-4982-9d4e-b674efda86b1" containerID="71b956a363ba4bc8cc2438b5fecd09e9f2b9e9e05efeca88f02e62fde2899617" exitCode=0 Oct 10 14:22:41 crc kubenswrapper[4745]: I1010 14:22:41.220634 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phgcv" event={"ID":"e91a3691-d1e1-4982-9d4e-b674efda86b1","Type":"ContainerDied","Data":"71b956a363ba4bc8cc2438b5fecd09e9f2b9e9e05efeca88f02e62fde2899617"} Oct 10 14:22:41 crc kubenswrapper[4745]: I1010 14:22:41.220866 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phgcv" event={"ID":"e91a3691-d1e1-4982-9d4e-b674efda86b1","Type":"ContainerStarted","Data":"2fa13700bba5079baa35aad11e20c75e19d400bddf633181a5998394af9ba154"} Oct 10 14:22:41 crc kubenswrapper[4745]: I1010 14:22:41.498009 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-fqb4s_839da6a0-320a-45d1-8a6c-be916ee27421/control-plane-machine-set-operator/0.log" Oct 10 14:22:41 crc kubenswrapper[4745]: I1010 14:22:41.669942 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-s8wbg_800c6118-3c93-4dd4-b7e4-17f23b82de1d/kube-rbac-proxy/0.log" Oct 10 14:22:41 crc kubenswrapper[4745]: I1010 14:22:41.759444 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-s8wbg_800c6118-3c93-4dd4-b7e4-17f23b82de1d/machine-api-operator/0.log" Oct 10 14:22:42 crc kubenswrapper[4745]: I1010 14:22:42.230836 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phgcv" event={"ID":"e91a3691-d1e1-4982-9d4e-b674efda86b1","Type":"ContainerStarted","Data":"a01ec12490ee3b185de9bc1c72a7f752d50bc5cd2361e1d1e79205900d7a42da"} Oct 10 14:22:43 crc kubenswrapper[4745]: I1010 14:22:43.241569 4745 generic.go:334] "Generic (PLEG): container finished" podID="e91a3691-d1e1-4982-9d4e-b674efda86b1" containerID="a01ec12490ee3b185de9bc1c72a7f752d50bc5cd2361e1d1e79205900d7a42da" exitCode=0 Oct 10 14:22:43 crc kubenswrapper[4745]: I1010 14:22:43.241634 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phgcv" event={"ID":"e91a3691-d1e1-4982-9d4e-b674efda86b1","Type":"ContainerDied","Data":"a01ec12490ee3b185de9bc1c72a7f752d50bc5cd2361e1d1e79205900d7a42da"} Oct 10 14:22:45 crc kubenswrapper[4745]: I1010 14:22:45.295944 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phgcv" event={"ID":"e91a3691-d1e1-4982-9d4e-b674efda86b1","Type":"ContainerStarted","Data":"307d9c716952201ea073c7e847ba9bd3925e1479403de39c230c7cb4b8a23d54"} Oct 10 14:22:45 crc kubenswrapper[4745]: I1010 14:22:45.323554 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-phgcv" podStartSLOduration=3.392950112 podStartE2EDuration="6.323534116s" podCreationTimestamp="2025-10-10 14:22:39 +0000 UTC" firstStartedPulling="2025-10-10 14:22:41.222328422 +0000 UTC m=+3875.119985195" lastFinishedPulling="2025-10-10 14:22:44.152912436 +0000 UTC m=+3878.050569199" observedRunningTime="2025-10-10 14:22:45.313500772 +0000 UTC m=+3879.211157555" watchObservedRunningTime="2025-10-10 14:22:45.323534116 +0000 UTC m=+3879.221190889" Oct 10 14:22:49 crc kubenswrapper[4745]: I1010 14:22:49.834488 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-phgcv" Oct 10 14:22:49 crc kubenswrapper[4745]: I1010 14:22:49.834972 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-phgcv" Oct 10 14:22:49 crc kubenswrapper[4745]: I1010 14:22:49.883631 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-phgcv" Oct 10 14:22:50 crc kubenswrapper[4745]: I1010 14:22:50.396615 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-phgcv" Oct 10 14:22:50 crc kubenswrapper[4745]: I1010 14:22:50.450040 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-phgcv"] Oct 10 14:22:52 crc kubenswrapper[4745]: I1010 14:22:52.356444 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-phgcv" podUID="e91a3691-d1e1-4982-9d4e-b674efda86b1" containerName="registry-server" containerID="cri-o://307d9c716952201ea073c7e847ba9bd3925e1479403de39c230c7cb4b8a23d54" gracePeriod=2 Oct 10 14:22:52 crc kubenswrapper[4745]: I1010 14:22:52.799391 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-phgcv" Oct 10 14:22:52 crc kubenswrapper[4745]: I1010 14:22:52.886554 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2ksj\" (UniqueName: \"kubernetes.io/projected/e91a3691-d1e1-4982-9d4e-b674efda86b1-kube-api-access-b2ksj\") pod \"e91a3691-d1e1-4982-9d4e-b674efda86b1\" (UID: \"e91a3691-d1e1-4982-9d4e-b674efda86b1\") " Oct 10 14:22:52 crc kubenswrapper[4745]: I1010 14:22:52.886630 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91a3691-d1e1-4982-9d4e-b674efda86b1-catalog-content\") pod \"e91a3691-d1e1-4982-9d4e-b674efda86b1\" (UID: \"e91a3691-d1e1-4982-9d4e-b674efda86b1\") " Oct 10 14:22:52 crc kubenswrapper[4745]: I1010 14:22:52.886864 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91a3691-d1e1-4982-9d4e-b674efda86b1-utilities\") pod \"e91a3691-d1e1-4982-9d4e-b674efda86b1\" (UID: \"e91a3691-d1e1-4982-9d4e-b674efda86b1\") " Oct 10 14:22:52 crc kubenswrapper[4745]: I1010 14:22:52.887901 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e91a3691-d1e1-4982-9d4e-b674efda86b1-utilities" (OuterVolumeSpecName: "utilities") pod "e91a3691-d1e1-4982-9d4e-b674efda86b1" (UID: "e91a3691-d1e1-4982-9d4e-b674efda86b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:22:52 crc kubenswrapper[4745]: I1010 14:22:52.891331 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e91a3691-d1e1-4982-9d4e-b674efda86b1-kube-api-access-b2ksj" (OuterVolumeSpecName: "kube-api-access-b2ksj") pod "e91a3691-d1e1-4982-9d4e-b674efda86b1" (UID: "e91a3691-d1e1-4982-9d4e-b674efda86b1"). InnerVolumeSpecName "kube-api-access-b2ksj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:22:52 crc kubenswrapper[4745]: I1010 14:22:52.978184 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e91a3691-d1e1-4982-9d4e-b674efda86b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e91a3691-d1e1-4982-9d4e-b674efda86b1" (UID: "e91a3691-d1e1-4982-9d4e-b674efda86b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:22:52 crc kubenswrapper[4745]: I1010 14:22:52.989542 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91a3691-d1e1-4982-9d4e-b674efda86b1-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:22:52 crc kubenswrapper[4745]: I1010 14:22:52.989575 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2ksj\" (UniqueName: \"kubernetes.io/projected/e91a3691-d1e1-4982-9d4e-b674efda86b1-kube-api-access-b2ksj\") on node \"crc\" DevicePath \"\"" Oct 10 14:22:52 crc kubenswrapper[4745]: I1010 14:22:52.989587 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91a3691-d1e1-4982-9d4e-b674efda86b1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:22:53 crc kubenswrapper[4745]: I1010 14:22:53.366026 4745 generic.go:334] "Generic (PLEG): container finished" podID="e91a3691-d1e1-4982-9d4e-b674efda86b1" containerID="307d9c716952201ea073c7e847ba9bd3925e1479403de39c230c7cb4b8a23d54" exitCode=0 Oct 10 14:22:53 crc kubenswrapper[4745]: I1010 14:22:53.366064 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phgcv" event={"ID":"e91a3691-d1e1-4982-9d4e-b674efda86b1","Type":"ContainerDied","Data":"307d9c716952201ea073c7e847ba9bd3925e1479403de39c230c7cb4b8a23d54"} Oct 10 14:22:53 crc kubenswrapper[4745]: I1010 14:22:53.366089 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phgcv" event={"ID":"e91a3691-d1e1-4982-9d4e-b674efda86b1","Type":"ContainerDied","Data":"2fa13700bba5079baa35aad11e20c75e19d400bddf633181a5998394af9ba154"} Oct 10 14:22:53 crc kubenswrapper[4745]: I1010 14:22:53.366104 4745 scope.go:117] "RemoveContainer" containerID="307d9c716952201ea073c7e847ba9bd3925e1479403de39c230c7cb4b8a23d54" Oct 10 14:22:53 crc kubenswrapper[4745]: I1010 14:22:53.366205 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-phgcv" Oct 10 14:22:53 crc kubenswrapper[4745]: I1010 14:22:53.390627 4745 scope.go:117] "RemoveContainer" containerID="a01ec12490ee3b185de9bc1c72a7f752d50bc5cd2361e1d1e79205900d7a42da" Oct 10 14:22:53 crc kubenswrapper[4745]: I1010 14:22:53.406918 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-phgcv"] Oct 10 14:22:53 crc kubenswrapper[4745]: I1010 14:22:53.414320 4745 scope.go:117] "RemoveContainer" containerID="71b956a363ba4bc8cc2438b5fecd09e9f2b9e9e05efeca88f02e62fde2899617" Oct 10 14:22:53 crc kubenswrapper[4745]: I1010 14:22:53.419823 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-phgcv"] Oct 10 14:22:53 crc kubenswrapper[4745]: I1010 14:22:53.514721 4745 scope.go:117] "RemoveContainer" containerID="307d9c716952201ea073c7e847ba9bd3925e1479403de39c230c7cb4b8a23d54" Oct 10 14:22:53 crc kubenswrapper[4745]: E1010 14:22:53.515273 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"307d9c716952201ea073c7e847ba9bd3925e1479403de39c230c7cb4b8a23d54\": container with ID starting with 307d9c716952201ea073c7e847ba9bd3925e1479403de39c230c7cb4b8a23d54 not found: ID does not exist" containerID="307d9c716952201ea073c7e847ba9bd3925e1479403de39c230c7cb4b8a23d54" Oct 10 14:22:53 crc kubenswrapper[4745]: I1010 14:22:53.515314 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"307d9c716952201ea073c7e847ba9bd3925e1479403de39c230c7cb4b8a23d54"} err="failed to get container status \"307d9c716952201ea073c7e847ba9bd3925e1479403de39c230c7cb4b8a23d54\": rpc error: code = NotFound desc = could not find container \"307d9c716952201ea073c7e847ba9bd3925e1479403de39c230c7cb4b8a23d54\": container with ID starting with 307d9c716952201ea073c7e847ba9bd3925e1479403de39c230c7cb4b8a23d54 not found: ID does not exist" Oct 10 14:22:53 crc kubenswrapper[4745]: I1010 14:22:53.515339 4745 scope.go:117] "RemoveContainer" containerID="a01ec12490ee3b185de9bc1c72a7f752d50bc5cd2361e1d1e79205900d7a42da" Oct 10 14:22:53 crc kubenswrapper[4745]: E1010 14:22:53.515595 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a01ec12490ee3b185de9bc1c72a7f752d50bc5cd2361e1d1e79205900d7a42da\": container with ID starting with a01ec12490ee3b185de9bc1c72a7f752d50bc5cd2361e1d1e79205900d7a42da not found: ID does not exist" containerID="a01ec12490ee3b185de9bc1c72a7f752d50bc5cd2361e1d1e79205900d7a42da" Oct 10 14:22:53 crc kubenswrapper[4745]: I1010 14:22:53.515624 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a01ec12490ee3b185de9bc1c72a7f752d50bc5cd2361e1d1e79205900d7a42da"} err="failed to get container status \"a01ec12490ee3b185de9bc1c72a7f752d50bc5cd2361e1d1e79205900d7a42da\": rpc error: code = NotFound desc = could not find container \"a01ec12490ee3b185de9bc1c72a7f752d50bc5cd2361e1d1e79205900d7a42da\": container with ID starting with a01ec12490ee3b185de9bc1c72a7f752d50bc5cd2361e1d1e79205900d7a42da not found: ID does not exist" Oct 10 14:22:53 crc kubenswrapper[4745]: I1010 14:22:53.515640 4745 scope.go:117] "RemoveContainer" containerID="71b956a363ba4bc8cc2438b5fecd09e9f2b9e9e05efeca88f02e62fde2899617" Oct 10 14:22:53 crc kubenswrapper[4745]: E1010 14:22:53.516017 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71b956a363ba4bc8cc2438b5fecd09e9f2b9e9e05efeca88f02e62fde2899617\": container with ID starting with 71b956a363ba4bc8cc2438b5fecd09e9f2b9e9e05efeca88f02e62fde2899617 not found: ID does not exist" containerID="71b956a363ba4bc8cc2438b5fecd09e9f2b9e9e05efeca88f02e62fde2899617" Oct 10 14:22:53 crc kubenswrapper[4745]: I1010 14:22:53.516037 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71b956a363ba4bc8cc2438b5fecd09e9f2b9e9e05efeca88f02e62fde2899617"} err="failed to get container status \"71b956a363ba4bc8cc2438b5fecd09e9f2b9e9e05efeca88f02e62fde2899617\": rpc error: code = NotFound desc = could not find container \"71b956a363ba4bc8cc2438b5fecd09e9f2b9e9e05efeca88f02e62fde2899617\": container with ID starting with 71b956a363ba4bc8cc2438b5fecd09e9f2b9e9e05efeca88f02e62fde2899617 not found: ID does not exist" Oct 10 14:22:54 crc kubenswrapper[4745]: I1010 14:22:54.263673 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-c24f6_61fb7cf8-4f9b-4414-824f-1f5270529840/cert-manager-controller/0.log" Oct 10 14:22:54 crc kubenswrapper[4745]: I1010 14:22:54.477585 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-rp25w_5f2820cb-2b97-4f6f-a92d-b6f16cb83de9/cert-manager-cainjector/0.log" Oct 10 14:22:54 crc kubenswrapper[4745]: I1010 14:22:54.486936 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-r4mqq_d365e19d-2015-4d4f-935e-79a4b3998dc4/cert-manager-webhook/0.log" Oct 10 14:22:54 crc kubenswrapper[4745]: I1010 14:22:54.755781 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e91a3691-d1e1-4982-9d4e-b674efda86b1" path="/var/lib/kubelet/pods/e91a3691-d1e1-4982-9d4e-b674efda86b1/volumes" Oct 10 14:23:06 crc kubenswrapper[4745]: I1010 14:23:06.109140 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-ws6j9_b8628d70-5b97-4c1c-b8c1-10e28a0439dd/nmstate-console-plugin/0.log" Oct 10 14:23:06 crc kubenswrapper[4745]: I1010 14:23:06.348370 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-bsftc_90c1fae7-ac4c-4d71-ac19-417d3f61f010/kube-rbac-proxy/0.log" Oct 10 14:23:06 crc kubenswrapper[4745]: I1010 14:23:06.350344 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-bsftc_90c1fae7-ac4c-4d71-ac19-417d3f61f010/nmstate-metrics/0.log" Oct 10 14:23:06 crc kubenswrapper[4745]: I1010 14:23:06.353079 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-6mcw2_9c18fb6e-a0ef-4367-a917-1abde904faf2/nmstate-handler/0.log" Oct 10 14:23:06 crc kubenswrapper[4745]: I1010 14:23:06.542552 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-8sw7s_f5837847-c246-4661-9174-b0ec21039982/nmstate-operator/0.log" Oct 10 14:23:06 crc kubenswrapper[4745]: I1010 14:23:06.591422 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-s6hqt_869f5cad-555c-4695-a9bc-f9e699484561/nmstate-webhook/0.log" Oct 10 14:23:20 crc kubenswrapper[4745]: I1010 14:23:20.822694 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-ktgdl_c1711da0-4710-4569-bdbd-2afe523b91be/kube-rbac-proxy/0.log" Oct 10 14:23:20 crc kubenswrapper[4745]: I1010 14:23:20.960217 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-ktgdl_c1711da0-4710-4569-bdbd-2afe523b91be/controller/0.log" Oct 10 14:23:21 crc kubenswrapper[4745]: I1010 14:23:21.028216 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-frr-files/0.log" Oct 10 14:23:21 crc kubenswrapper[4745]: I1010 14:23:21.173847 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-reloader/0.log" Oct 10 14:23:21 crc kubenswrapper[4745]: I1010 14:23:21.197314 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-metrics/0.log" Oct 10 14:23:21 crc kubenswrapper[4745]: I1010 14:23:21.218682 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-frr-files/0.log" Oct 10 14:23:21 crc kubenswrapper[4745]: I1010 14:23:21.257839 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-reloader/0.log" Oct 10 14:23:21 crc kubenswrapper[4745]: I1010 14:23:21.460804 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-frr-files/0.log" Oct 10 14:23:21 crc kubenswrapper[4745]: I1010 14:23:21.466265 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-metrics/0.log" Oct 10 14:23:21 crc kubenswrapper[4745]: I1010 14:23:21.521999 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-metrics/0.log" Oct 10 14:23:21 crc kubenswrapper[4745]: I1010 14:23:21.544329 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-reloader/0.log" Oct 10 14:23:21 crc kubenswrapper[4745]: I1010 14:23:21.904462 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-frr-files/0.log" Oct 10 14:23:21 crc kubenswrapper[4745]: I1010 14:23:21.925037 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-metrics/0.log" Oct 10 14:23:21 crc kubenswrapper[4745]: I1010 14:23:21.939171 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/controller/0.log" Oct 10 14:23:21 crc kubenswrapper[4745]: I1010 14:23:21.940613 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/cp-reloader/0.log" Oct 10 14:23:22 crc kubenswrapper[4745]: I1010 14:23:22.160927 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/kube-rbac-proxy/0.log" Oct 10 14:23:22 crc kubenswrapper[4745]: I1010 14:23:22.167486 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/kube-rbac-proxy-frr/0.log" Oct 10 14:23:22 crc kubenswrapper[4745]: I1010 14:23:22.178297 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/frr-metrics/0.log" Oct 10 14:23:22 crc kubenswrapper[4745]: I1010 14:23:22.404514 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/reloader/0.log" Oct 10 14:23:22 crc kubenswrapper[4745]: I1010 14:23:22.414018 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-8ktnf_a362b7b7-7534-4a4f-ae57-aac84fd292c7/frr-k8s-webhook-server/0.log" Oct 10 14:23:22 crc kubenswrapper[4745]: I1010 14:23:22.647014 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6d67b47897-wgz9q_bdf86da2-3d6e-44f6-bd60-1ddfad7c6ff1/manager/0.log" Oct 10 14:23:22 crc kubenswrapper[4745]: I1010 14:23:22.856571 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-55796c598f-6xdbw_fcb7295e-30a8-4023-9f78-c75a0dadea04/webhook-server/0.log" Oct 10 14:23:22 crc kubenswrapper[4745]: I1010 14:23:22.988958 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-92z9w_b298d9ff-9b10-42f4-aabf-259673933a1a/kube-rbac-proxy/0.log" Oct 10 14:23:23 crc kubenswrapper[4745]: I1010 14:23:23.336884 4745 scope.go:117] "RemoveContainer" containerID="e1137d9abefc625ad35fb8e19592cf3efd2aef854d7954080a4810dedf3e9616" Oct 10 14:23:23 crc kubenswrapper[4745]: I1010 14:23:23.541188 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-52wk2_e3feebb0-e7f8-4aef-a725-2b9cf003364c/frr/0.log" Oct 10 14:23:23 crc kubenswrapper[4745]: I1010 14:23:23.551807 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-92z9w_b298d9ff-9b10-42f4-aabf-259673933a1a/speaker/0.log" Oct 10 14:23:34 crc kubenswrapper[4745]: I1010 14:23:34.785185 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl_2050d745-31cf-44f8-8fb8-e1a4701c9799/util/0.log" Oct 10 14:23:34 crc kubenswrapper[4745]: I1010 14:23:34.973398 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl_2050d745-31cf-44f8-8fb8-e1a4701c9799/util/0.log" Oct 10 14:23:34 crc kubenswrapper[4745]: I1010 14:23:34.990192 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl_2050d745-31cf-44f8-8fb8-e1a4701c9799/pull/0.log" Oct 10 14:23:35 crc kubenswrapper[4745]: I1010 14:23:35.017587 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl_2050d745-31cf-44f8-8fb8-e1a4701c9799/pull/0.log" Oct 10 14:23:35 crc kubenswrapper[4745]: I1010 14:23:35.168121 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl_2050d745-31cf-44f8-8fb8-e1a4701c9799/pull/0.log" Oct 10 14:23:35 crc kubenswrapper[4745]: I1010 14:23:35.212245 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl_2050d745-31cf-44f8-8fb8-e1a4701c9799/util/0.log" Oct 10 14:23:35 crc kubenswrapper[4745]: I1010 14:23:35.223420 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frfvl_2050d745-31cf-44f8-8fb8-e1a4701c9799/extract/0.log" Oct 10 14:23:35 crc kubenswrapper[4745]: I1010 14:23:35.361369 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h5nbz_78f0f7b2-023b-4227-a64a-c3ccae73349b/extract-utilities/0.log" Oct 10 14:23:35 crc kubenswrapper[4745]: I1010 14:23:35.595795 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h5nbz_78f0f7b2-023b-4227-a64a-c3ccae73349b/extract-utilities/0.log" Oct 10 14:23:35 crc kubenswrapper[4745]: I1010 14:23:35.614008 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h5nbz_78f0f7b2-023b-4227-a64a-c3ccae73349b/extract-content/0.log" Oct 10 14:23:35 crc kubenswrapper[4745]: I1010 14:23:35.614141 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h5nbz_78f0f7b2-023b-4227-a64a-c3ccae73349b/extract-content/0.log" Oct 10 14:23:35 crc kubenswrapper[4745]: I1010 14:23:35.777309 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h5nbz_78f0f7b2-023b-4227-a64a-c3ccae73349b/extract-utilities/0.log" Oct 10 14:23:35 crc kubenswrapper[4745]: I1010 14:23:35.787275 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h5nbz_78f0f7b2-023b-4227-a64a-c3ccae73349b/extract-content/0.log" Oct 10 14:23:35 crc kubenswrapper[4745]: I1010 14:23:35.979078 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-stdb8_ba2a7571-2a26-4bf9-a642-ef927ed9c7f8/extract-utilities/0.log" Oct 10 14:23:36 crc kubenswrapper[4745]: I1010 14:23:36.255786 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-stdb8_ba2a7571-2a26-4bf9-a642-ef927ed9c7f8/extract-content/0.log" Oct 10 14:23:36 crc kubenswrapper[4745]: I1010 14:23:36.293353 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-stdb8_ba2a7571-2a26-4bf9-a642-ef927ed9c7f8/extract-utilities/0.log" Oct 10 14:23:36 crc kubenswrapper[4745]: I1010 14:23:36.304325 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h5nbz_78f0f7b2-023b-4227-a64a-c3ccae73349b/registry-server/0.log" Oct 10 14:23:36 crc kubenswrapper[4745]: I1010 14:23:36.341146 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-stdb8_ba2a7571-2a26-4bf9-a642-ef927ed9c7f8/extract-content/0.log" Oct 10 14:23:36 crc kubenswrapper[4745]: I1010 14:23:36.442403 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-stdb8_ba2a7571-2a26-4bf9-a642-ef927ed9c7f8/extract-utilities/0.log" Oct 10 14:23:36 crc kubenswrapper[4745]: I1010 14:23:36.453419 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-stdb8_ba2a7571-2a26-4bf9-a642-ef927ed9c7f8/extract-content/0.log" Oct 10 14:23:36 crc kubenswrapper[4745]: I1010 14:23:36.666795 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c_4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4/util/0.log" Oct 10 14:23:36 crc kubenswrapper[4745]: I1010 14:23:36.854985 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-stdb8_ba2a7571-2a26-4bf9-a642-ef927ed9c7f8/registry-server/0.log" Oct 10 14:23:36 crc kubenswrapper[4745]: I1010 14:23:36.908641 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c_4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4/util/0.log" Oct 10 14:23:36 crc kubenswrapper[4745]: I1010 14:23:36.909189 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c_4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4/pull/0.log" Oct 10 14:23:36 crc kubenswrapper[4745]: I1010 14:23:36.937908 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c_4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4/pull/0.log" Oct 10 14:23:37 crc kubenswrapper[4745]: I1010 14:23:37.049789 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c_4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4/util/0.log" Oct 10 14:23:37 crc kubenswrapper[4745]: I1010 14:23:37.075639 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c_4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4/pull/0.log" Oct 10 14:23:37 crc kubenswrapper[4745]: I1010 14:23:37.092150 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjnk6c_4cc0cf48-9860-42b1-ad6c-cbb6559ad8d4/extract/0.log" Oct 10 14:23:37 crc kubenswrapper[4745]: I1010 14:23:37.206703 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hbrh4_a87ec403-812a-4af1-83eb-95788143fcc4/marketplace-operator/0.log" Oct 10 14:23:37 crc kubenswrapper[4745]: I1010 14:23:37.262683 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bwpkj_511552f9-6c48-4742-9332-a5724f4d5697/extract-utilities/0.log" Oct 10 14:23:37 crc kubenswrapper[4745]: I1010 14:23:37.425848 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bwpkj_511552f9-6c48-4742-9332-a5724f4d5697/extract-content/0.log" Oct 10 14:23:37 crc kubenswrapper[4745]: I1010 14:23:37.442775 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bwpkj_511552f9-6c48-4742-9332-a5724f4d5697/extract-utilities/0.log" Oct 10 14:23:37 crc kubenswrapper[4745]: I1010 14:23:37.455462 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bwpkj_511552f9-6c48-4742-9332-a5724f4d5697/extract-content/0.log" Oct 10 14:23:37 crc kubenswrapper[4745]: I1010 14:23:37.605522 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bwpkj_511552f9-6c48-4742-9332-a5724f4d5697/extract-content/0.log" Oct 10 14:23:37 crc kubenswrapper[4745]: I1010 14:23:37.607514 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bwpkj_511552f9-6c48-4742-9332-a5724f4d5697/extract-utilities/0.log" Oct 10 14:23:37 crc kubenswrapper[4745]: I1010 14:23:37.732932 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bwpkj_511552f9-6c48-4742-9332-a5724f4d5697/registry-server/0.log" Oct 10 14:23:37 crc kubenswrapper[4745]: I1010 14:23:37.836787 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k7bwk_c3284f0e-0f23-4b38-bc27-3b2806155073/extract-utilities/0.log" Oct 10 14:23:37 crc kubenswrapper[4745]: I1010 14:23:37.980352 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k7bwk_c3284f0e-0f23-4b38-bc27-3b2806155073/extract-utilities/0.log" Oct 10 14:23:37 crc kubenswrapper[4745]: I1010 14:23:37.986700 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k7bwk_c3284f0e-0f23-4b38-bc27-3b2806155073/extract-content/0.log" Oct 10 14:23:37 crc kubenswrapper[4745]: I1010 14:23:37.999153 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k7bwk_c3284f0e-0f23-4b38-bc27-3b2806155073/extract-content/0.log" Oct 10 14:23:38 crc kubenswrapper[4745]: I1010 14:23:38.177112 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k7bwk_c3284f0e-0f23-4b38-bc27-3b2806155073/extract-utilities/0.log" Oct 10 14:23:38 crc kubenswrapper[4745]: I1010 14:23:38.181769 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k7bwk_c3284f0e-0f23-4b38-bc27-3b2806155073/extract-content/0.log" Oct 10 14:23:38 crc kubenswrapper[4745]: I1010 14:23:38.754056 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k7bwk_c3284f0e-0f23-4b38-bc27-3b2806155073/registry-server/0.log" Oct 10 14:24:16 crc kubenswrapper[4745]: I1010 14:24:16.187552 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:24:16 crc kubenswrapper[4745]: I1010 14:24:16.188239 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:24:46 crc kubenswrapper[4745]: I1010 14:24:46.186528 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:24:46 crc kubenswrapper[4745]: I1010 14:24:46.187291 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:25:09 crc kubenswrapper[4745]: I1010 14:25:09.428275 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5l4xw"] Oct 10 14:25:09 crc kubenswrapper[4745]: E1010 14:25:09.429184 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e91a3691-d1e1-4982-9d4e-b674efda86b1" containerName="registry-server" Oct 10 14:25:09 crc kubenswrapper[4745]: I1010 14:25:09.429473 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="e91a3691-d1e1-4982-9d4e-b674efda86b1" containerName="registry-server" Oct 10 14:25:09 crc kubenswrapper[4745]: E1010 14:25:09.429487 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e91a3691-d1e1-4982-9d4e-b674efda86b1" containerName="extract-content" Oct 10 14:25:09 crc kubenswrapper[4745]: I1010 14:25:09.429494 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="e91a3691-d1e1-4982-9d4e-b674efda86b1" containerName="extract-content" Oct 10 14:25:09 crc kubenswrapper[4745]: E1010 14:25:09.429532 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e91a3691-d1e1-4982-9d4e-b674efda86b1" containerName="extract-utilities" Oct 10 14:25:09 crc kubenswrapper[4745]: I1010 14:25:09.429787 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="e91a3691-d1e1-4982-9d4e-b674efda86b1" containerName="extract-utilities" Oct 10 14:25:09 crc kubenswrapper[4745]: I1010 14:25:09.430000 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="e91a3691-d1e1-4982-9d4e-b674efda86b1" containerName="registry-server" Oct 10 14:25:09 crc kubenswrapper[4745]: I1010 14:25:09.431888 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5l4xw" Oct 10 14:25:09 crc kubenswrapper[4745]: I1010 14:25:09.449065 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5l4xw"] Oct 10 14:25:09 crc kubenswrapper[4745]: I1010 14:25:09.512775 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcqsk\" (UniqueName: \"kubernetes.io/projected/97306df9-f9ae-40da-ac39-e84c7483f0b8-kube-api-access-kcqsk\") pod \"certified-operators-5l4xw\" (UID: \"97306df9-f9ae-40da-ac39-e84c7483f0b8\") " pod="openshift-marketplace/certified-operators-5l4xw" Oct 10 14:25:09 crc kubenswrapper[4745]: I1010 14:25:09.513433 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97306df9-f9ae-40da-ac39-e84c7483f0b8-catalog-content\") pod \"certified-operators-5l4xw\" (UID: \"97306df9-f9ae-40da-ac39-e84c7483f0b8\") " pod="openshift-marketplace/certified-operators-5l4xw" Oct 10 14:25:09 crc kubenswrapper[4745]: I1010 14:25:09.513473 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97306df9-f9ae-40da-ac39-e84c7483f0b8-utilities\") pod \"certified-operators-5l4xw\" (UID: \"97306df9-f9ae-40da-ac39-e84c7483f0b8\") " pod="openshift-marketplace/certified-operators-5l4xw" Oct 10 14:25:09 crc kubenswrapper[4745]: I1010 14:25:09.615158 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcqsk\" (UniqueName: \"kubernetes.io/projected/97306df9-f9ae-40da-ac39-e84c7483f0b8-kube-api-access-kcqsk\") pod \"certified-operators-5l4xw\" (UID: \"97306df9-f9ae-40da-ac39-e84c7483f0b8\") " pod="openshift-marketplace/certified-operators-5l4xw" Oct 10 14:25:09 crc kubenswrapper[4745]: I1010 14:25:09.615235 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97306df9-f9ae-40da-ac39-e84c7483f0b8-catalog-content\") pod \"certified-operators-5l4xw\" (UID: \"97306df9-f9ae-40da-ac39-e84c7483f0b8\") " pod="openshift-marketplace/certified-operators-5l4xw" Oct 10 14:25:09 crc kubenswrapper[4745]: I1010 14:25:09.615260 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97306df9-f9ae-40da-ac39-e84c7483f0b8-utilities\") pod \"certified-operators-5l4xw\" (UID: \"97306df9-f9ae-40da-ac39-e84c7483f0b8\") " pod="openshift-marketplace/certified-operators-5l4xw" Oct 10 14:25:09 crc kubenswrapper[4745]: I1010 14:25:09.615841 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97306df9-f9ae-40da-ac39-e84c7483f0b8-catalog-content\") pod \"certified-operators-5l4xw\" (UID: \"97306df9-f9ae-40da-ac39-e84c7483f0b8\") " pod="openshift-marketplace/certified-operators-5l4xw" Oct 10 14:25:09 crc kubenswrapper[4745]: I1010 14:25:09.615869 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97306df9-f9ae-40da-ac39-e84c7483f0b8-utilities\") pod \"certified-operators-5l4xw\" (UID: \"97306df9-f9ae-40da-ac39-e84c7483f0b8\") " pod="openshift-marketplace/certified-operators-5l4xw" Oct 10 14:25:09 crc kubenswrapper[4745]: I1010 14:25:09.645035 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcqsk\" (UniqueName: \"kubernetes.io/projected/97306df9-f9ae-40da-ac39-e84c7483f0b8-kube-api-access-kcqsk\") pod \"certified-operators-5l4xw\" (UID: \"97306df9-f9ae-40da-ac39-e84c7483f0b8\") " pod="openshift-marketplace/certified-operators-5l4xw" Oct 10 14:25:09 crc kubenswrapper[4745]: I1010 14:25:09.817488 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5l4xw" Oct 10 14:25:10 crc kubenswrapper[4745]: I1010 14:25:10.293243 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5l4xw"] Oct 10 14:25:10 crc kubenswrapper[4745]: I1010 14:25:10.637012 4745 generic.go:334] "Generic (PLEG): container finished" podID="97306df9-f9ae-40da-ac39-e84c7483f0b8" containerID="40f360dd3aaa67deb5e9a4e3dc276435667b2856f4b241b59370230e96f3828b" exitCode=0 Oct 10 14:25:10 crc kubenswrapper[4745]: I1010 14:25:10.637084 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5l4xw" event={"ID":"97306df9-f9ae-40da-ac39-e84c7483f0b8","Type":"ContainerDied","Data":"40f360dd3aaa67deb5e9a4e3dc276435667b2856f4b241b59370230e96f3828b"} Oct 10 14:25:10 crc kubenswrapper[4745]: I1010 14:25:10.637993 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5l4xw" event={"ID":"97306df9-f9ae-40da-ac39-e84c7483f0b8","Type":"ContainerStarted","Data":"ef65f62cb2c2e2068e9c4a618b7c4fd2057cf51737f679777b60dbff66e21b6f"} Oct 10 14:25:12 crc kubenswrapper[4745]: I1010 14:25:12.654587 4745 generic.go:334] "Generic (PLEG): container finished" podID="97306df9-f9ae-40da-ac39-e84c7483f0b8" containerID="7d113958786b5405298731feefaa50bbb2b1f780474d5d56a2bcb5f27d0cb0f0" exitCode=0 Oct 10 14:25:12 crc kubenswrapper[4745]: I1010 14:25:12.654794 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5l4xw" event={"ID":"97306df9-f9ae-40da-ac39-e84c7483f0b8","Type":"ContainerDied","Data":"7d113958786b5405298731feefaa50bbb2b1f780474d5d56a2bcb5f27d0cb0f0"} Oct 10 14:25:13 crc kubenswrapper[4745]: I1010 14:25:13.666327 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5l4xw" event={"ID":"97306df9-f9ae-40da-ac39-e84c7483f0b8","Type":"ContainerStarted","Data":"d1c58c1c9be3a9a50c2a861485c2754b4c33f1dd3822959107ed16f600c51ae9"} Oct 10 14:25:13 crc kubenswrapper[4745]: I1010 14:25:13.685375 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5l4xw" podStartSLOduration=2.213211166 podStartE2EDuration="4.685354431s" podCreationTimestamp="2025-10-10 14:25:09 +0000 UTC" firstStartedPulling="2025-10-10 14:25:10.638423031 +0000 UTC m=+4024.536079794" lastFinishedPulling="2025-10-10 14:25:13.110566266 +0000 UTC m=+4027.008223059" observedRunningTime="2025-10-10 14:25:13.684666764 +0000 UTC m=+4027.582323537" watchObservedRunningTime="2025-10-10 14:25:13.685354431 +0000 UTC m=+4027.583011194" Oct 10 14:25:16 crc kubenswrapper[4745]: I1010 14:25:16.186881 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:25:16 crc kubenswrapper[4745]: I1010 14:25:16.187566 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:25:16 crc kubenswrapper[4745]: I1010 14:25:16.187632 4745 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" Oct 10 14:25:16 crc kubenswrapper[4745]: I1010 14:25:16.188783 4745 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e4dc030f0d245a23b97a0415e5a9289fcfa1518fa939913ff0ab29c429c1984"} pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 14:25:16 crc kubenswrapper[4745]: I1010 14:25:16.188881 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" containerID="cri-o://4e4dc030f0d245a23b97a0415e5a9289fcfa1518fa939913ff0ab29c429c1984" gracePeriod=600 Oct 10 14:25:16 crc kubenswrapper[4745]: I1010 14:25:16.695956 4745 generic.go:334] "Generic (PLEG): container finished" podID="19294049-0379-4bd4-bdaa-c46d50052501" containerID="3b94cccace5d99f025668855489c519fe976bed9dca92d6d811f8f458152e094" exitCode=0 Oct 10 14:25:16 crc kubenswrapper[4745]: I1010 14:25:16.696047 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mns4j/must-gather-6c92c" event={"ID":"19294049-0379-4bd4-bdaa-c46d50052501","Type":"ContainerDied","Data":"3b94cccace5d99f025668855489c519fe976bed9dca92d6d811f8f458152e094"} Oct 10 14:25:16 crc kubenswrapper[4745]: I1010 14:25:16.696711 4745 scope.go:117] "RemoveContainer" containerID="3b94cccace5d99f025668855489c519fe976bed9dca92d6d811f8f458152e094" Oct 10 14:25:16 crc kubenswrapper[4745]: I1010 14:25:16.701147 4745 generic.go:334] "Generic (PLEG): container finished" podID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerID="4e4dc030f0d245a23b97a0415e5a9289fcfa1518fa939913ff0ab29c429c1984" exitCode=0 Oct 10 14:25:16 crc kubenswrapper[4745]: I1010 14:25:16.701193 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerDied","Data":"4e4dc030f0d245a23b97a0415e5a9289fcfa1518fa939913ff0ab29c429c1984"} Oct 10 14:25:16 crc kubenswrapper[4745]: I1010 14:25:16.701228 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" event={"ID":"5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c","Type":"ContainerStarted","Data":"9dd4efa37ab5bf6b14c73e26c9bebe9ed2e27077917ca78739641af303fefa1f"} Oct 10 14:25:16 crc kubenswrapper[4745]: I1010 14:25:16.701245 4745 scope.go:117] "RemoveContainer" containerID="a6ac9a2ef7d49730766ff1ac064d0fe7ad967d67c770352a2a823b5c13bf3264" Oct 10 14:25:17 crc kubenswrapper[4745]: I1010 14:25:17.114637 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mns4j_must-gather-6c92c_19294049-0379-4bd4-bdaa-c46d50052501/gather/0.log" Oct 10 14:25:19 crc kubenswrapper[4745]: I1010 14:25:19.817905 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5l4xw" Oct 10 14:25:19 crc kubenswrapper[4745]: I1010 14:25:19.820979 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5l4xw" Oct 10 14:25:19 crc kubenswrapper[4745]: I1010 14:25:19.869190 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5l4xw" Oct 10 14:25:20 crc kubenswrapper[4745]: I1010 14:25:20.844669 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5l4xw" Oct 10 14:25:20 crc kubenswrapper[4745]: I1010 14:25:20.890541 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5l4xw"] Oct 10 14:25:22 crc kubenswrapper[4745]: I1010 14:25:22.774112 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5l4xw" podUID="97306df9-f9ae-40da-ac39-e84c7483f0b8" containerName="registry-server" containerID="cri-o://d1c58c1c9be3a9a50c2a861485c2754b4c33f1dd3822959107ed16f600c51ae9" gracePeriod=2 Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.263264 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5l4xw" Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.415425 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97306df9-f9ae-40da-ac39-e84c7483f0b8-catalog-content\") pod \"97306df9-f9ae-40da-ac39-e84c7483f0b8\" (UID: \"97306df9-f9ae-40da-ac39-e84c7483f0b8\") " Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.415505 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97306df9-f9ae-40da-ac39-e84c7483f0b8-utilities\") pod \"97306df9-f9ae-40da-ac39-e84c7483f0b8\" (UID: \"97306df9-f9ae-40da-ac39-e84c7483f0b8\") " Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.415634 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcqsk\" (UniqueName: \"kubernetes.io/projected/97306df9-f9ae-40da-ac39-e84c7483f0b8-kube-api-access-kcqsk\") pod \"97306df9-f9ae-40da-ac39-e84c7483f0b8\" (UID: \"97306df9-f9ae-40da-ac39-e84c7483f0b8\") " Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.417124 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97306df9-f9ae-40da-ac39-e84c7483f0b8-utilities" (OuterVolumeSpecName: "utilities") pod "97306df9-f9ae-40da-ac39-e84c7483f0b8" (UID: "97306df9-f9ae-40da-ac39-e84c7483f0b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.421973 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97306df9-f9ae-40da-ac39-e84c7483f0b8-kube-api-access-kcqsk" (OuterVolumeSpecName: "kube-api-access-kcqsk") pod "97306df9-f9ae-40da-ac39-e84c7483f0b8" (UID: "97306df9-f9ae-40da-ac39-e84c7483f0b8"). InnerVolumeSpecName "kube-api-access-kcqsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.463998 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97306df9-f9ae-40da-ac39-e84c7483f0b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97306df9-f9ae-40da-ac39-e84c7483f0b8" (UID: "97306df9-f9ae-40da-ac39-e84c7483f0b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.517888 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97306df9-f9ae-40da-ac39-e84c7483f0b8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.517920 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97306df9-f9ae-40da-ac39-e84c7483f0b8-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.517932 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcqsk\" (UniqueName: \"kubernetes.io/projected/97306df9-f9ae-40da-ac39-e84c7483f0b8-kube-api-access-kcqsk\") on node \"crc\" DevicePath \"\"" Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.787702 4745 generic.go:334] "Generic (PLEG): container finished" podID="97306df9-f9ae-40da-ac39-e84c7483f0b8" containerID="d1c58c1c9be3a9a50c2a861485c2754b4c33f1dd3822959107ed16f600c51ae9" exitCode=0 Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.787768 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5l4xw" event={"ID":"97306df9-f9ae-40da-ac39-e84c7483f0b8","Type":"ContainerDied","Data":"d1c58c1c9be3a9a50c2a861485c2754b4c33f1dd3822959107ed16f600c51ae9"} Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.787854 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5l4xw" Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.788052 4745 scope.go:117] "RemoveContainer" containerID="d1c58c1c9be3a9a50c2a861485c2754b4c33f1dd3822959107ed16f600c51ae9" Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.788034 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5l4xw" event={"ID":"97306df9-f9ae-40da-ac39-e84c7483f0b8","Type":"ContainerDied","Data":"ef65f62cb2c2e2068e9c4a618b7c4fd2057cf51737f679777b60dbff66e21b6f"} Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.814534 4745 scope.go:117] "RemoveContainer" containerID="7d113958786b5405298731feefaa50bbb2b1f780474d5d56a2bcb5f27d0cb0f0" Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.841509 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5l4xw"] Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.855713 4745 scope.go:117] "RemoveContainer" containerID="40f360dd3aaa67deb5e9a4e3dc276435667b2856f4b241b59370230e96f3828b" Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.856891 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5l4xw"] Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.893595 4745 scope.go:117] "RemoveContainer" containerID="d1c58c1c9be3a9a50c2a861485c2754b4c33f1dd3822959107ed16f600c51ae9" Oct 10 14:25:23 crc kubenswrapper[4745]: E1010 14:25:23.894610 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1c58c1c9be3a9a50c2a861485c2754b4c33f1dd3822959107ed16f600c51ae9\": container with ID starting with d1c58c1c9be3a9a50c2a861485c2754b4c33f1dd3822959107ed16f600c51ae9 not found: ID does not exist" containerID="d1c58c1c9be3a9a50c2a861485c2754b4c33f1dd3822959107ed16f600c51ae9" Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.894673 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1c58c1c9be3a9a50c2a861485c2754b4c33f1dd3822959107ed16f600c51ae9"} err="failed to get container status \"d1c58c1c9be3a9a50c2a861485c2754b4c33f1dd3822959107ed16f600c51ae9\": rpc error: code = NotFound desc = could not find container \"d1c58c1c9be3a9a50c2a861485c2754b4c33f1dd3822959107ed16f600c51ae9\": container with ID starting with d1c58c1c9be3a9a50c2a861485c2754b4c33f1dd3822959107ed16f600c51ae9 not found: ID does not exist" Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.894695 4745 scope.go:117] "RemoveContainer" containerID="7d113958786b5405298731feefaa50bbb2b1f780474d5d56a2bcb5f27d0cb0f0" Oct 10 14:25:23 crc kubenswrapper[4745]: E1010 14:25:23.895334 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d113958786b5405298731feefaa50bbb2b1f780474d5d56a2bcb5f27d0cb0f0\": container with ID starting with 7d113958786b5405298731feefaa50bbb2b1f780474d5d56a2bcb5f27d0cb0f0 not found: ID does not exist" containerID="7d113958786b5405298731feefaa50bbb2b1f780474d5d56a2bcb5f27d0cb0f0" Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.895357 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d113958786b5405298731feefaa50bbb2b1f780474d5d56a2bcb5f27d0cb0f0"} err="failed to get container status \"7d113958786b5405298731feefaa50bbb2b1f780474d5d56a2bcb5f27d0cb0f0\": rpc error: code = NotFound desc = could not find container \"7d113958786b5405298731feefaa50bbb2b1f780474d5d56a2bcb5f27d0cb0f0\": container with ID starting with 7d113958786b5405298731feefaa50bbb2b1f780474d5d56a2bcb5f27d0cb0f0 not found: ID does not exist" Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.895371 4745 scope.go:117] "RemoveContainer" containerID="40f360dd3aaa67deb5e9a4e3dc276435667b2856f4b241b59370230e96f3828b" Oct 10 14:25:23 crc kubenswrapper[4745]: E1010 14:25:23.895707 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40f360dd3aaa67deb5e9a4e3dc276435667b2856f4b241b59370230e96f3828b\": container with ID starting with 40f360dd3aaa67deb5e9a4e3dc276435667b2856f4b241b59370230e96f3828b not found: ID does not exist" containerID="40f360dd3aaa67deb5e9a4e3dc276435667b2856f4b241b59370230e96f3828b" Oct 10 14:25:23 crc kubenswrapper[4745]: I1010 14:25:23.895769 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40f360dd3aaa67deb5e9a4e3dc276435667b2856f4b241b59370230e96f3828b"} err="failed to get container status \"40f360dd3aaa67deb5e9a4e3dc276435667b2856f4b241b59370230e96f3828b\": rpc error: code = NotFound desc = could not find container \"40f360dd3aaa67deb5e9a4e3dc276435667b2856f4b241b59370230e96f3828b\": container with ID starting with 40f360dd3aaa67deb5e9a4e3dc276435667b2856f4b241b59370230e96f3828b not found: ID does not exist" Oct 10 14:25:24 crc kubenswrapper[4745]: I1010 14:25:24.756183 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97306df9-f9ae-40da-ac39-e84c7483f0b8" path="/var/lib/kubelet/pods/97306df9-f9ae-40da-ac39-e84c7483f0b8/volumes" Oct 10 14:25:26 crc kubenswrapper[4745]: I1010 14:25:26.713892 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mns4j/must-gather-6c92c"] Oct 10 14:25:26 crc kubenswrapper[4745]: I1010 14:25:26.715644 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mns4j/must-gather-6c92c" podUID="19294049-0379-4bd4-bdaa-c46d50052501" containerName="copy" containerID="cri-o://a69f0f0c0f239785f5fc334c8faa6c6d1ee6660f38b299d5917c040e1dfa9347" gracePeriod=2 Oct 10 14:25:26 crc kubenswrapper[4745]: I1010 14:25:26.724867 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mns4j/must-gather-6c92c"] Oct 10 14:25:27 crc kubenswrapper[4745]: I1010 14:25:27.136442 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mns4j_must-gather-6c92c_19294049-0379-4bd4-bdaa-c46d50052501/copy/0.log" Oct 10 14:25:27 crc kubenswrapper[4745]: I1010 14:25:27.137688 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mns4j/must-gather-6c92c" Oct 10 14:25:27 crc kubenswrapper[4745]: I1010 14:25:27.292911 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvpjj\" (UniqueName: \"kubernetes.io/projected/19294049-0379-4bd4-bdaa-c46d50052501-kube-api-access-bvpjj\") pod \"19294049-0379-4bd4-bdaa-c46d50052501\" (UID: \"19294049-0379-4bd4-bdaa-c46d50052501\") " Oct 10 14:25:27 crc kubenswrapper[4745]: I1010 14:25:27.293028 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/19294049-0379-4bd4-bdaa-c46d50052501-must-gather-output\") pod \"19294049-0379-4bd4-bdaa-c46d50052501\" (UID: \"19294049-0379-4bd4-bdaa-c46d50052501\") " Oct 10 14:25:27 crc kubenswrapper[4745]: I1010 14:25:27.298192 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19294049-0379-4bd4-bdaa-c46d50052501-kube-api-access-bvpjj" (OuterVolumeSpecName: "kube-api-access-bvpjj") pod "19294049-0379-4bd4-bdaa-c46d50052501" (UID: "19294049-0379-4bd4-bdaa-c46d50052501"). InnerVolumeSpecName "kube-api-access-bvpjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:25:27 crc kubenswrapper[4745]: I1010 14:25:27.395471 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvpjj\" (UniqueName: \"kubernetes.io/projected/19294049-0379-4bd4-bdaa-c46d50052501-kube-api-access-bvpjj\") on node \"crc\" DevicePath \"\"" Oct 10 14:25:27 crc kubenswrapper[4745]: I1010 14:25:27.458124 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19294049-0379-4bd4-bdaa-c46d50052501-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "19294049-0379-4bd4-bdaa-c46d50052501" (UID: "19294049-0379-4bd4-bdaa-c46d50052501"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:25:27 crc kubenswrapper[4745]: I1010 14:25:27.497337 4745 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/19294049-0379-4bd4-bdaa-c46d50052501-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 10 14:25:27 crc kubenswrapper[4745]: I1010 14:25:27.827706 4745 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mns4j_must-gather-6c92c_19294049-0379-4bd4-bdaa-c46d50052501/copy/0.log" Oct 10 14:25:27 crc kubenswrapper[4745]: I1010 14:25:27.828060 4745 generic.go:334] "Generic (PLEG): container finished" podID="19294049-0379-4bd4-bdaa-c46d50052501" containerID="a69f0f0c0f239785f5fc334c8faa6c6d1ee6660f38b299d5917c040e1dfa9347" exitCode=143 Oct 10 14:25:27 crc kubenswrapper[4745]: I1010 14:25:27.828105 4745 scope.go:117] "RemoveContainer" containerID="a69f0f0c0f239785f5fc334c8faa6c6d1ee6660f38b299d5917c040e1dfa9347" Oct 10 14:25:27 crc kubenswrapper[4745]: I1010 14:25:27.828148 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mns4j/must-gather-6c92c" Oct 10 14:25:27 crc kubenswrapper[4745]: I1010 14:25:27.845183 4745 scope.go:117] "RemoveContainer" containerID="3b94cccace5d99f025668855489c519fe976bed9dca92d6d811f8f458152e094" Oct 10 14:25:28 crc kubenswrapper[4745]: I1010 14:25:28.179471 4745 scope.go:117] "RemoveContainer" containerID="a69f0f0c0f239785f5fc334c8faa6c6d1ee6660f38b299d5917c040e1dfa9347" Oct 10 14:25:28 crc kubenswrapper[4745]: E1010 14:25:28.179898 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a69f0f0c0f239785f5fc334c8faa6c6d1ee6660f38b299d5917c040e1dfa9347\": container with ID starting with a69f0f0c0f239785f5fc334c8faa6c6d1ee6660f38b299d5917c040e1dfa9347 not found: ID does not exist" containerID="a69f0f0c0f239785f5fc334c8faa6c6d1ee6660f38b299d5917c040e1dfa9347" Oct 10 14:25:28 crc kubenswrapper[4745]: I1010 14:25:28.179929 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a69f0f0c0f239785f5fc334c8faa6c6d1ee6660f38b299d5917c040e1dfa9347"} err="failed to get container status \"a69f0f0c0f239785f5fc334c8faa6c6d1ee6660f38b299d5917c040e1dfa9347\": rpc error: code = NotFound desc = could not find container \"a69f0f0c0f239785f5fc334c8faa6c6d1ee6660f38b299d5917c040e1dfa9347\": container with ID starting with a69f0f0c0f239785f5fc334c8faa6c6d1ee6660f38b299d5917c040e1dfa9347 not found: ID does not exist" Oct 10 14:25:28 crc kubenswrapper[4745]: I1010 14:25:28.179953 4745 scope.go:117] "RemoveContainer" containerID="3b94cccace5d99f025668855489c519fe976bed9dca92d6d811f8f458152e094" Oct 10 14:25:28 crc kubenswrapper[4745]: E1010 14:25:28.180290 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b94cccace5d99f025668855489c519fe976bed9dca92d6d811f8f458152e094\": container with ID starting with 3b94cccace5d99f025668855489c519fe976bed9dca92d6d811f8f458152e094 not found: ID does not exist" containerID="3b94cccace5d99f025668855489c519fe976bed9dca92d6d811f8f458152e094" Oct 10 14:25:28 crc kubenswrapper[4745]: I1010 14:25:28.180309 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b94cccace5d99f025668855489c519fe976bed9dca92d6d811f8f458152e094"} err="failed to get container status \"3b94cccace5d99f025668855489c519fe976bed9dca92d6d811f8f458152e094\": rpc error: code = NotFound desc = could not find container \"3b94cccace5d99f025668855489c519fe976bed9dca92d6d811f8f458152e094\": container with ID starting with 3b94cccace5d99f025668855489c519fe976bed9dca92d6d811f8f458152e094 not found: ID does not exist" Oct 10 14:25:28 crc kubenswrapper[4745]: I1010 14:25:28.762196 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19294049-0379-4bd4-bdaa-c46d50052501" path="/var/lib/kubelet/pods/19294049-0379-4bd4-bdaa-c46d50052501/volumes" Oct 10 14:26:09 crc kubenswrapper[4745]: I1010 14:26:09.779332 4745 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xvc7s"] Oct 10 14:26:09 crc kubenswrapper[4745]: E1010 14:26:09.780194 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19294049-0379-4bd4-bdaa-c46d50052501" containerName="gather" Oct 10 14:26:09 crc kubenswrapper[4745]: I1010 14:26:09.780205 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="19294049-0379-4bd4-bdaa-c46d50052501" containerName="gather" Oct 10 14:26:09 crc kubenswrapper[4745]: E1010 14:26:09.780228 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19294049-0379-4bd4-bdaa-c46d50052501" containerName="copy" Oct 10 14:26:09 crc kubenswrapper[4745]: I1010 14:26:09.780233 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="19294049-0379-4bd4-bdaa-c46d50052501" containerName="copy" Oct 10 14:26:09 crc kubenswrapper[4745]: E1010 14:26:09.780240 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97306df9-f9ae-40da-ac39-e84c7483f0b8" containerName="registry-server" Oct 10 14:26:09 crc kubenswrapper[4745]: I1010 14:26:09.780246 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="97306df9-f9ae-40da-ac39-e84c7483f0b8" containerName="registry-server" Oct 10 14:26:09 crc kubenswrapper[4745]: E1010 14:26:09.780257 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97306df9-f9ae-40da-ac39-e84c7483f0b8" containerName="extract-content" Oct 10 14:26:09 crc kubenswrapper[4745]: I1010 14:26:09.780263 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="97306df9-f9ae-40da-ac39-e84c7483f0b8" containerName="extract-content" Oct 10 14:26:09 crc kubenswrapper[4745]: E1010 14:26:09.780274 4745 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97306df9-f9ae-40da-ac39-e84c7483f0b8" containerName="extract-utilities" Oct 10 14:26:09 crc kubenswrapper[4745]: I1010 14:26:09.780281 4745 state_mem.go:107] "Deleted CPUSet assignment" podUID="97306df9-f9ae-40da-ac39-e84c7483f0b8" containerName="extract-utilities" Oct 10 14:26:09 crc kubenswrapper[4745]: I1010 14:26:09.780446 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="19294049-0379-4bd4-bdaa-c46d50052501" containerName="gather" Oct 10 14:26:09 crc kubenswrapper[4745]: I1010 14:26:09.780471 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="97306df9-f9ae-40da-ac39-e84c7483f0b8" containerName="registry-server" Oct 10 14:26:09 crc kubenswrapper[4745]: I1010 14:26:09.780484 4745 memory_manager.go:354] "RemoveStaleState removing state" podUID="19294049-0379-4bd4-bdaa-c46d50052501" containerName="copy" Oct 10 14:26:09 crc kubenswrapper[4745]: I1010 14:26:09.781765 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvc7s" Oct 10 14:26:09 crc kubenswrapper[4745]: I1010 14:26:09.800005 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xvc7s"] Oct 10 14:26:09 crc kubenswrapper[4745]: I1010 14:26:09.911340 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-utilities\") pod \"community-operators-xvc7s\" (UID: \"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad\") " pod="openshift-marketplace/community-operators-xvc7s" Oct 10 14:26:09 crc kubenswrapper[4745]: I1010 14:26:09.912522 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttr99\" (UniqueName: \"kubernetes.io/projected/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-kube-api-access-ttr99\") pod \"community-operators-xvc7s\" (UID: \"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad\") " pod="openshift-marketplace/community-operators-xvc7s" Oct 10 14:26:09 crc kubenswrapper[4745]: I1010 14:26:09.912708 4745 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-catalog-content\") pod \"community-operators-xvc7s\" (UID: \"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad\") " pod="openshift-marketplace/community-operators-xvc7s" Oct 10 14:26:10 crc kubenswrapper[4745]: I1010 14:26:10.014392 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-utilities\") pod \"community-operators-xvc7s\" (UID: \"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad\") " pod="openshift-marketplace/community-operators-xvc7s" Oct 10 14:26:10 crc kubenswrapper[4745]: I1010 14:26:10.014479 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttr99\" (UniqueName: \"kubernetes.io/projected/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-kube-api-access-ttr99\") pod \"community-operators-xvc7s\" (UID: \"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad\") " pod="openshift-marketplace/community-operators-xvc7s" Oct 10 14:26:10 crc kubenswrapper[4745]: I1010 14:26:10.014519 4745 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-catalog-content\") pod \"community-operators-xvc7s\" (UID: \"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad\") " pod="openshift-marketplace/community-operators-xvc7s" Oct 10 14:26:10 crc kubenswrapper[4745]: I1010 14:26:10.014946 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-catalog-content\") pod \"community-operators-xvc7s\" (UID: \"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad\") " pod="openshift-marketplace/community-operators-xvc7s" Oct 10 14:26:10 crc kubenswrapper[4745]: I1010 14:26:10.015002 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-utilities\") pod \"community-operators-xvc7s\" (UID: \"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad\") " pod="openshift-marketplace/community-operators-xvc7s" Oct 10 14:26:10 crc kubenswrapper[4745]: I1010 14:26:10.038196 4745 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttr99\" (UniqueName: \"kubernetes.io/projected/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-kube-api-access-ttr99\") pod \"community-operators-xvc7s\" (UID: \"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad\") " pod="openshift-marketplace/community-operators-xvc7s" Oct 10 14:26:10 crc kubenswrapper[4745]: I1010 14:26:10.101688 4745 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvc7s" Oct 10 14:26:10 crc kubenswrapper[4745]: I1010 14:26:10.638843 4745 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xvc7s"] Oct 10 14:26:11 crc kubenswrapper[4745]: I1010 14:26:11.282956 4745 generic.go:334] "Generic (PLEG): container finished" podID="38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad" containerID="6bb57befff056bd3c0cdfb4c4f6eef3fe5ee664b2dfe360c61d7c232448d4049" exitCode=0 Oct 10 14:26:11 crc kubenswrapper[4745]: I1010 14:26:11.283012 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvc7s" event={"ID":"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad","Type":"ContainerDied","Data":"6bb57befff056bd3c0cdfb4c4f6eef3fe5ee664b2dfe360c61d7c232448d4049"} Oct 10 14:26:11 crc kubenswrapper[4745]: I1010 14:26:11.283316 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvc7s" event={"ID":"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad","Type":"ContainerStarted","Data":"86e56db489ba1e797f5962dd002201590f5b4850eecfdac3960880774b02ebe3"} Oct 10 14:26:13 crc kubenswrapper[4745]: I1010 14:26:13.299464 4745 generic.go:334] "Generic (PLEG): container finished" podID="38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad" containerID="ce054a3db25ac1b332a7cf081e2d6fd65d4d5ff3f960265a4c3c6a40619c3ded" exitCode=0 Oct 10 14:26:13 crc kubenswrapper[4745]: I1010 14:26:13.299528 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvc7s" event={"ID":"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad","Type":"ContainerDied","Data":"ce054a3db25ac1b332a7cf081e2d6fd65d4d5ff3f960265a4c3c6a40619c3ded"} Oct 10 14:26:15 crc kubenswrapper[4745]: I1010 14:26:15.319826 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvc7s" event={"ID":"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad","Type":"ContainerStarted","Data":"4f139d832209fa6ab97f6d0f7c92bc868815926c743b11c23699a805727a78ef"} Oct 10 14:26:15 crc kubenswrapper[4745]: I1010 14:26:15.337289 4745 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xvc7s" podStartSLOduration=3.4322591510000002 podStartE2EDuration="6.337272089s" podCreationTimestamp="2025-10-10 14:26:09 +0000 UTC" firstStartedPulling="2025-10-10 14:26:11.285117952 +0000 UTC m=+4085.182774715" lastFinishedPulling="2025-10-10 14:26:14.19013089 +0000 UTC m=+4088.087787653" observedRunningTime="2025-10-10 14:26:15.333895265 +0000 UTC m=+4089.231552038" watchObservedRunningTime="2025-10-10 14:26:15.337272089 +0000 UTC m=+4089.234928852" Oct 10 14:26:20 crc kubenswrapper[4745]: I1010 14:26:20.101997 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xvc7s" Oct 10 14:26:20 crc kubenswrapper[4745]: I1010 14:26:20.102556 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xvc7s" Oct 10 14:26:20 crc kubenswrapper[4745]: I1010 14:26:20.153016 4745 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xvc7s" Oct 10 14:26:20 crc kubenswrapper[4745]: I1010 14:26:20.424173 4745 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xvc7s" Oct 10 14:26:20 crc kubenswrapper[4745]: I1010 14:26:20.472675 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xvc7s"] Oct 10 14:26:22 crc kubenswrapper[4745]: I1010 14:26:22.395270 4745 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xvc7s" podUID="38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad" containerName="registry-server" containerID="cri-o://4f139d832209fa6ab97f6d0f7c92bc868815926c743b11c23699a805727a78ef" gracePeriod=2 Oct 10 14:26:22 crc kubenswrapper[4745]: I1010 14:26:22.908571 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvc7s" Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.058412 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-utilities\") pod \"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad\" (UID: \"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad\") " Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.058554 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-catalog-content\") pod \"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad\" (UID: \"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad\") " Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.058600 4745 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttr99\" (UniqueName: \"kubernetes.io/projected/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-kube-api-access-ttr99\") pod \"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad\" (UID: \"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad\") " Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.064151 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-kube-api-access-ttr99" (OuterVolumeSpecName: "kube-api-access-ttr99") pod "38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad" (UID: "38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad"). InnerVolumeSpecName "kube-api-access-ttr99". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.070961 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-utilities" (OuterVolumeSpecName: "utilities") pod "38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad" (UID: "38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.112096 4745 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad" (UID: "38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.161121 4745 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.161429 4745 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttr99\" (UniqueName: \"kubernetes.io/projected/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-kube-api-access-ttr99\") on node \"crc\" DevicePath \"\"" Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.161439 4745 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.404780 4745 generic.go:334] "Generic (PLEG): container finished" podID="38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad" containerID="4f139d832209fa6ab97f6d0f7c92bc868815926c743b11c23699a805727a78ef" exitCode=0 Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.404821 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvc7s" event={"ID":"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad","Type":"ContainerDied","Data":"4f139d832209fa6ab97f6d0f7c92bc868815926c743b11c23699a805727a78ef"} Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.404846 4745 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvc7s" event={"ID":"38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad","Type":"ContainerDied","Data":"86e56db489ba1e797f5962dd002201590f5b4850eecfdac3960880774b02ebe3"} Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.404862 4745 scope.go:117] "RemoveContainer" containerID="4f139d832209fa6ab97f6d0f7c92bc868815926c743b11c23699a805727a78ef" Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.404975 4745 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvc7s" Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.428724 4745 scope.go:117] "RemoveContainer" containerID="ce054a3db25ac1b332a7cf081e2d6fd65d4d5ff3f960265a4c3c6a40619c3ded" Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.446798 4745 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xvc7s"] Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.457788 4745 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xvc7s"] Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.634653 4745 scope.go:117] "RemoveContainer" containerID="6bb57befff056bd3c0cdfb4c4f6eef3fe5ee664b2dfe360c61d7c232448d4049" Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.704195 4745 scope.go:117] "RemoveContainer" containerID="4f139d832209fa6ab97f6d0f7c92bc868815926c743b11c23699a805727a78ef" Oct 10 14:26:23 crc kubenswrapper[4745]: E1010 14:26:23.704678 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f139d832209fa6ab97f6d0f7c92bc868815926c743b11c23699a805727a78ef\": container with ID starting with 4f139d832209fa6ab97f6d0f7c92bc868815926c743b11c23699a805727a78ef not found: ID does not exist" containerID="4f139d832209fa6ab97f6d0f7c92bc868815926c743b11c23699a805727a78ef" Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.704709 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f139d832209fa6ab97f6d0f7c92bc868815926c743b11c23699a805727a78ef"} err="failed to get container status \"4f139d832209fa6ab97f6d0f7c92bc868815926c743b11c23699a805727a78ef\": rpc error: code = NotFound desc = could not find container \"4f139d832209fa6ab97f6d0f7c92bc868815926c743b11c23699a805727a78ef\": container with ID starting with 4f139d832209fa6ab97f6d0f7c92bc868815926c743b11c23699a805727a78ef not found: ID does not exist" Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.704741 4745 scope.go:117] "RemoveContainer" containerID="ce054a3db25ac1b332a7cf081e2d6fd65d4d5ff3f960265a4c3c6a40619c3ded" Oct 10 14:26:23 crc kubenswrapper[4745]: E1010 14:26:23.705152 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce054a3db25ac1b332a7cf081e2d6fd65d4d5ff3f960265a4c3c6a40619c3ded\": container with ID starting with ce054a3db25ac1b332a7cf081e2d6fd65d4d5ff3f960265a4c3c6a40619c3ded not found: ID does not exist" containerID="ce054a3db25ac1b332a7cf081e2d6fd65d4d5ff3f960265a4c3c6a40619c3ded" Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.705174 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce054a3db25ac1b332a7cf081e2d6fd65d4d5ff3f960265a4c3c6a40619c3ded"} err="failed to get container status \"ce054a3db25ac1b332a7cf081e2d6fd65d4d5ff3f960265a4c3c6a40619c3ded\": rpc error: code = NotFound desc = could not find container \"ce054a3db25ac1b332a7cf081e2d6fd65d4d5ff3f960265a4c3c6a40619c3ded\": container with ID starting with ce054a3db25ac1b332a7cf081e2d6fd65d4d5ff3f960265a4c3c6a40619c3ded not found: ID does not exist" Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.705189 4745 scope.go:117] "RemoveContainer" containerID="6bb57befff056bd3c0cdfb4c4f6eef3fe5ee664b2dfe360c61d7c232448d4049" Oct 10 14:26:23 crc kubenswrapper[4745]: E1010 14:26:23.705456 4745 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bb57befff056bd3c0cdfb4c4f6eef3fe5ee664b2dfe360c61d7c232448d4049\": container with ID starting with 6bb57befff056bd3c0cdfb4c4f6eef3fe5ee664b2dfe360c61d7c232448d4049 not found: ID does not exist" containerID="6bb57befff056bd3c0cdfb4c4f6eef3fe5ee664b2dfe360c61d7c232448d4049" Oct 10 14:26:23 crc kubenswrapper[4745]: I1010 14:26:23.705480 4745 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bb57befff056bd3c0cdfb4c4f6eef3fe5ee664b2dfe360c61d7c232448d4049"} err="failed to get container status \"6bb57befff056bd3c0cdfb4c4f6eef3fe5ee664b2dfe360c61d7c232448d4049\": rpc error: code = NotFound desc = could not find container \"6bb57befff056bd3c0cdfb4c4f6eef3fe5ee664b2dfe360c61d7c232448d4049\": container with ID starting with 6bb57befff056bd3c0cdfb4c4f6eef3fe5ee664b2dfe360c61d7c232448d4049 not found: ID does not exist" Oct 10 14:26:24 crc kubenswrapper[4745]: I1010 14:26:24.756000 4745 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad" path="/var/lib/kubelet/pods/38ecb8b7-e909-4fb1-bbd2-788dfa0d4fad/volumes" Oct 10 14:27:16 crc kubenswrapper[4745]: I1010 14:27:16.186920 4745 patch_prober.go:28] interesting pod/machine-config-daemon-4dnq2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:27:16 crc kubenswrapper[4745]: I1010 14:27:16.187461 4745 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dnq2" podUID="5aaf7b57-cecd-4c6c-a45b-74504fa2dc2c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:27:23 crc kubenswrapper[4745]: I1010 14:27:23.682464 4745 scope.go:117] "RemoveContainer" containerID="389a1b800d641a9b9abf749b5854bd3ce0e7f9617e7653944ac38cd5c77448f7" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515072213737024454 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015072213740017363 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015072203307016504 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015072203307015454 5ustar corecore